var/home/core/zuul-output/0000755000175000017500000000000015113677114014533 5ustar corecorevar/home/core/zuul-output/logs/0000755000175000017500000000000015113704232015467 5ustar corecorevar/home/core/zuul-output/logs/kubelet.log0000644000000000000000004014364415113704223017701 0ustar rootrootDec 03 00:06:50 crc systemd[1]: Starting Kubernetes Kubelet... Dec 03 00:06:50 crc restorecon[4698]: Relabeled /var/lib/kubelet/config.json from system_u:object_r:unlabeled_t:s0 to system_u:object_r:container_var_lib_t:s0 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/device-plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/device-plugins/kubelet.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/volumes/kubernetes.io~configmap/nginx-conf/..2025_02_23_05_40_35.4114275528/nginx.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/22e96971 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/21c98286 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/0f1869e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/46889d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/5b6a5969 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/6c7921f5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4804f443 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/2a46b283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/a6b5573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4f88ee5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/5a4eee4b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/cd87c521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/38602af4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/1483b002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/0346718b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/d3ed4ada not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/3bb473a5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/8cd075a9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/00ab4760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/54a21c09 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/70478888 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/43802770 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/955a0edc not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/bca2d009 not reset as customized by admin to system_u:object_r:container_file_t:s0:c140,c1009 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/b295f9bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/bc46ea27 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5731fc1b not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5e1b2a3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/943f0936 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/3f764ee4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/8695e3f9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/aed7aa86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/c64d7448 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/0ba16bd2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/207a939f not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/54aa8cdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/1f5fa595 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/bf9c8153 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/47fba4ea not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/7ae55ce9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7906a268 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/ce43fa69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7fc7ea3a not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/d8c38b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/9ef015fb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/b9db6a41 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/b1733d79 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/afccd338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/9df0a185 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/18938cf8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/7ab4eb23 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/56930be6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_35.630010865 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/0d8e3722 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/d22b2e76 not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/e036759f not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/2734c483 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/57878fe7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/3f3c2e58 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/375bec3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/7bc41e08 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/48c7a72d not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/4b66701f not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/a5a1c202 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_40.1388695756 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/26f3df5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/6d8fb21d not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/50e94777 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208473b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/ec9e08ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3b787c39 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208eaed5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/93aa3a2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3c697968 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/ba950ec9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/cb5cdb37 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/f2df9827 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/fedaa673 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/9ca2df95 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/b2d7460e not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2207853c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/241c1c29 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2d910eaf not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/c6c0f2e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/399edc97 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8049f7cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/0cec5484 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/312446d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c406,c828 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8e56a35d not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/2d30ddb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/eca8053d not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/c3a25c9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c168,c522 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/b9609c22 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/e8b0eca9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/b36a9c3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/38af7b07 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/ae821620 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/baa23338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/2c534809 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/59b29eae not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/c91a8e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/4d87494a not reset as customized by admin to system_u:object_r:container_file_t:s0:c442,c857 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/1e33ca63 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/8dea7be2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d0b04a99 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d84f01e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/4109059b not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/a7258a3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/05bdf2b6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/f3261b51 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/315d045e not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/5fdcf278 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/d053f757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/c2850dc7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fcfb0b2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c7ac9b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fa0c0d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c609b6ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/2be6c296 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/89a32653 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/4eb9afeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/13af6efa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/b03f9724 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/e3d105cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/3aed4d83 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/0765fa6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/2cefc627 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/3dcc6345 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/365af391 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b1130c0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/236a5913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b9432e26 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/5ddb0e3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/986dc4fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/8a23ff9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/9728ae68 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/665f31d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/136c9b42 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/98a1575b not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/cac69136 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/5deb77a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/2ae53400 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/e46f2326 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/dc688d3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/3497c3cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/177eb008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/af5a2afa not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/d780cb1f not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/49b0f374 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/26fbb125 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/cf14125a not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/b7f86972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/e51d739c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/88ba6a69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/669a9acf not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/5cd51231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/75349ec7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/15c26839 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/45023dcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/2bb66a50 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/64d03bdd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/ab8e7ca0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/bb9be25f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/9a0b61d3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/d471b9d2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/8cb76b8e not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/11a00840 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/ec355a92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/992f735e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d59cdbbc not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/72133ff0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/c56c834c not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d13724c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/0a498258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa471982 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fc900d92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa7d68da not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/4bacf9b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/424021b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/fc2e31a3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/f51eefac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/c8997f2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/7481f599 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/fdafea19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/d0e1c571 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/ee398915 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/682bb6b8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a3e67855 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a989f289 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/915431bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/7796fdab not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/dcdb5f19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/a3aaa88c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/5508e3e6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/160585de not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/e99f8da3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/8bc85570 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/a5861c91 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/84db1135 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/9e1a6043 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/c1aba1c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/d55ccd6d not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/971cc9f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/8f2e3dcf not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/ceb35e9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/1c192745 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/5209e501 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/f83de4df not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/e7b978ac not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/c64304a1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/5384386b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/cce3e3ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/8fb75465 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/740f573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/32fd1134 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/0a861bd3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/80363026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/bfa952a8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..2025_02_23_05_33_31.333075221 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/793bf43d not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/7db1bb6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/4f6a0368 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/c12c7d86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/36c4a773 not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/4c1e98ae not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/a4c8115c not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/setup/7db1802e not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver/a008a7ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-syncer/2c836bac not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-regeneration-controller/0ce62299 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-insecure-readyz/945d2457 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-check-endpoints/7d5c1dd8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/index.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/bundle-v1.15.0.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/channel.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/package.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/bc8d0691 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/6b76097a not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/34d1af30 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/312ba61c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/645d5dd1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/16e825f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/4cf51fc9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/2a23d348 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/075dbd49 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/dd585ddd not reset as customized by admin to system_u:object_r:container_file_t:s0:c377,c642 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/17ebd0ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c343 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/005579f4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_23_11.1287037894 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/bf5f3b9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/af276eb7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/ea28e322 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/692e6683 not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/871746a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/4eb2e958 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/ca9b62da not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/0edd6fce not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/containers/controller-manager/89b4555f not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/655fcd71 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/0d43c002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/e68efd17 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/9acf9b65 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/5ae3ff11 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/1e59206a not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/27af16d1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c304,c1017 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/7918e729 not reset as customized by admin to system_u:object_r:container_file_t:s0:c853,c893 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/5d976d0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c585,c981 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/d7f55cbb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/f0812073 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/1a56cbeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/7fdd437e not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/cdfb5652 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/fix-audit-permissions/fb93119e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver/f1e8fc0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver-check-endpoints/218511f3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server/serving-certs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/ca8af7b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/72cc8a75 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/6e8a3760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4c3455c0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/2278acb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4b453e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/3ec09bda not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2/cacerts.bin not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java/cacerts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl/ca-bundle.trust.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/email-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/objsign-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2ae6433e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fde84897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75680d2e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/openshift-service-serving-signer_1740288168.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/facfc4fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f5a969c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CFCA_EV_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9ef4a08a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ingress-operator_1740288202.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2f332aed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/248c8271.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d10a21f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ACCVRAIZ1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a94d09e5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c9a4d3b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40193066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd8c0d63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b936d1c6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CA_Disig_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4fd49c6c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM_SERVIDORES_SEGUROS.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b81b93f0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f9a69fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b30d5fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ANF_Secure_Server_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b433981b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93851c9e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9282e51c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7dd1bc4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Actalis_Authentication_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/930ac5d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f47b495.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e113c810.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5931b5bc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Commercial.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2b349938.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e48193cf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/302904dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a716d4ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Networking.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93bc0acc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/86212b19.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b727005e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbc54cab.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f51bb24c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c28a8a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9c8dfbd4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ccc52f49.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cb1c3204.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ce5e74ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd08c599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6d41d539.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb5fa911.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e35234b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8cb5ee0f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a7c655d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f8fc53da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/de6d66f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d41b5e2a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/41a3f684.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1df5a75f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_2011.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e36a6752.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b872f2b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9576d26b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/228f89db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_ECC_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb717492.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d21b73c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b1b94ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/595e996b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_RSA_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b46e03d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/128f4b91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_3_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81f2d2b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Autoridad_de_Certificacion_Firmaprofesional_CIF_A62634068.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3bde41ac.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d16a5865.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_EC-384_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0179095f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ffa7f1eb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9482e63a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4dae3dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e359ba6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7e067d03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/95aff9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7746a63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Baltimore_CyberTrust_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/653b494a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3ad48a91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_2_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/54657681.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/82223c44.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8de2f56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d9dafe4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d96b65e2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee64a828.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40547a79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5a3f0ff8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a780d93.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/34d996fb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/eed8c118.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/89c02a45.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b1159c4c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d6325660.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4c339cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8312c4c1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_E1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8508e720.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5fdd185d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48bec511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/69105f4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b9bc432.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/32888f65.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b03dec0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/219d9499.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5acf816d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbf06781.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc99f41e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AAA_Certificate_Services.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/985c1f52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8794b4e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_BR_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7c037b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ef954a4e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_EV_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2add47b6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/90c5a3c8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0f3e76e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/53a1b57a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_EV_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5ad8a5d6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/68dd7389.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d04f354.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d6437c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/062cdee6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bd43e1dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7f3d5d1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c491639e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3513523f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/399e7759.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/feffd413.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d18e9066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/607986c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c90bc37d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1b0f7e5c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e08bfd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dd8e9d41.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed39abd0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a3418fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bc3f2570.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_High_Assurance_EV_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/244b5494.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81b9768f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4be590e0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_ECC_P384_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9846683b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/252252d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e8e7201.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_RSA4096_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d52c538d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c44cc0c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Trusted_Root_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75d1b2ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a2c66da8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ecccd8db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust.net_Certification_Authority__2048_.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/aee5f10d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e7271e8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0e59380.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4c3982f2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b99d060.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf64f35b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0a775a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/002c0b4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cc450945.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_EC1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/106f3e4d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b3fb433b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4042bcee.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/02265526.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/455f1b52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0d69c7e1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9f727ac7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5e98733a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0cd152c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc4d6a89.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6187b673.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/FIRMAPROFESIONAL_CA_ROOT-A_WEB.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ba8887ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/068570d1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f081611a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48a195d8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GDCA_TrustAUTH_R5_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f6fa695.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab59055e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b92fd57f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GLOBALTRUST_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fa5da96b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ec40989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7719f463.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1001acf7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f013ecaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/626dceaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c559d742.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1d3472b9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9479c8c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a81e292b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4bfab552.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e071171e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/57bcb2da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_ECC_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab5346f4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5046c355.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_RSA_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/865fbdf9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da0cfd1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/85cde254.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_ECC_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbb3f32b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureSign_RootCA11.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5860aaa6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/31188b5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HiPKI_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c7f1359b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f15c80c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hongkong_Post_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/09789157.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/18856ac4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e09d511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Commercial_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cf701eeb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d06393bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Public_Sector_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/10531352.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Izenpe.com.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureTrust_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0ed035a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsec_e-Szigno_Root_CA_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8160b96c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8651083.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2c63f966.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_ECC_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d89cda1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/01419da9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_RSA_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7a5b843.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_RSA_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf53fb88.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9591a472.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3afde786.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Gold_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NAVER_Global_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3fb36b73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d39b0a2c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a89d74c2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd58d51e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7db1890.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NetLock_Arany__Class_Gold__F__tan__s__tv__ny.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/988a38cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/60afe812.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f39fc864.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5443e9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GB_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e73d606e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dfc0fe80.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b66938e9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e1eab7c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GC_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/773e07ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c899c73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d59297b8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ddcda989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_1_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/749e9e03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/52b525c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7e8dc79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a819ef2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/08063a00.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b483515.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/064e0aa9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1f58a078.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6f7454b3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7fa05551.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76faf6c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9339512a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f387163d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee37c333.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e18bfb83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e442e424.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fe8a2cd8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/23f4c490.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5cd81ad7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0c70a8d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7892ad52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SZAFIR_ROOT_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4f316efb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_RSA_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/06dc52d5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/583d0756.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0bf05006.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/88950faa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9046744a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c860d51.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_RSA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6fa5da56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/33ee480d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Secure_Global_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/63a2c897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_ECC_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bdacca6f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ff34af3f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbff3a01.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_ECC_RootCA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_C1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/406c9bb1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_C3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Services_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Silver_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/99e1b953.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/14bc7599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TUBITAK_Kamu_SM_SSL_Kok_Sertifikasi_-_Surum_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a3adc42.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f459871d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_ECC_Root_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_RSA_Root_2023.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TeliaSonera_Root_CA_v1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telia_Root_CA_v2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f103249.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f058632f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-certificates.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9bf03295.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/98aaf404.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1cef98f5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/073bfcc5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2923b3f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f249de83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/edcbddb5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P256_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b5697b0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ae85e5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b74d2bd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P384_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d887a5bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9aef356c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TunTrust_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd64f3fc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e13665f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Extended_Validation_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f5dc4f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da7377f6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Global_G2_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c01eb047.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/304d27c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed858448.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f30dd6ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/04f60c28.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_ECC_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fc5a8f99.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/35105088.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee532fd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/XRamp_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/706f604c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76579174.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d86cdd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/882de061.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f618aec.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a9d40e02.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e-Szigno_Root_CA_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e868b802.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/83e9984f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ePKI_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca6e4ad9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d6523ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4b718d9b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/869fbf79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/containers/registry/f8d22bdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/6e8bbfac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/54dd7996 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/a4f1bb05 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/207129da not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/c1df39e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/15b8f1cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/77bd6913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/2382c1b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/704ce128 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/70d16fe0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/bfb95535 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/57a8e8e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/1b9d3e5e not reset as customized by admin to system_u:object_r:container_file_t:s0:c107,c917 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/fddb173c not reset as customized by admin to system_u:object_r:container_file_t:s0:c202,c983 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/95d3c6c4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/bfb5fff5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/2aef40aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/c0391cad not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/1119e69d not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/660608b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/8220bd53 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/85f99d5c not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/4b0225f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/9c2a3394 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/e820b243 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/1ca52ea0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/e6988e45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/6655f00b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/98bc3986 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/08e3458a not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/2a191cb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/6c4eeefb not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/f61a549c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/24891863 not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/fbdfd89c not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/9b63b3bc not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/8acde6d6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/node-driver-registrar/59ecbba3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/csi-provisioner/685d4be3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/containers/route-controller-manager/feaea55e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/63709497 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/d966b7fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/f5773757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/81c9edb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/57bf57ee not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/86f5e6aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/0aabe31d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/d2af85c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/09d157d9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c0fe7256 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c30319e4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/e6b1dd45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/2bb643f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/920de426 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/70fa1e87 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/a1c12a2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/9442e6c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/5b45ec72 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/3c9f3a59 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/1091c11b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/9a6821c6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/ec0c35e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/517f37e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/6214fe78 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/ba189c8b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/351e4f31 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/c0f219ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/8069f607 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/559c3d82 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/605ad488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/148df488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/3bf6dcb4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/022a2feb not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/938c3924 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/729fe23e not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/1fd5cbd4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/a96697e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/e155ddca not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/10dd0e0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/6f2c8392 not reset as customized by admin to system_u:object_r:container_file_t:s0:c267,c588 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/bd241ad9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/plugins/csi-hostpath not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/plugins/csi-hostpath/csi.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/plugins/kubernetes.io not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/plugins/kubernetes.io/csi not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983 not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/vol_data.json not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 03 00:06:50 crc restorecon[4698]: /var/lib/kubelet/plugins_registry not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 03 00:06:50 crc restorecon[4698]: Relabeled /var/usrlocal/bin/kubenswrapper from system_u:object_r:bin_t:s0 to system_u:object_r:kubelet_exec_t:s0 Dec 03 00:06:51 crc kubenswrapper[4953]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Dec 03 00:06:51 crc kubenswrapper[4953]: Flag --minimum-container-ttl-duration has been deprecated, Use --eviction-hard or --eviction-soft instead. Will be removed in a future version. Dec 03 00:06:51 crc kubenswrapper[4953]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Dec 03 00:06:51 crc kubenswrapper[4953]: Flag --register-with-taints has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Dec 03 00:06:51 crc kubenswrapper[4953]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Dec 03 00:06:51 crc kubenswrapper[4953]: Flag --system-reserved has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.160576 4953 server.go:211] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Dec 03 00:06:51 crc kubenswrapper[4953]: W1203 00:06:51.163772 4953 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Dec 03 00:06:51 crc kubenswrapper[4953]: W1203 00:06:51.163795 4953 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Dec 03 00:06:51 crc kubenswrapper[4953]: W1203 00:06:51.163801 4953 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Dec 03 00:06:51 crc kubenswrapper[4953]: W1203 00:06:51.163807 4953 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Dec 03 00:06:51 crc kubenswrapper[4953]: W1203 00:06:51.163812 4953 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Dec 03 00:06:51 crc kubenswrapper[4953]: W1203 00:06:51.163818 4953 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Dec 03 00:06:51 crc kubenswrapper[4953]: W1203 00:06:51.163823 4953 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Dec 03 00:06:51 crc kubenswrapper[4953]: W1203 00:06:51.163829 4953 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Dec 03 00:06:51 crc kubenswrapper[4953]: W1203 00:06:51.163836 4953 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Dec 03 00:06:51 crc kubenswrapper[4953]: W1203 00:06:51.163842 4953 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Dec 03 00:06:51 crc kubenswrapper[4953]: W1203 00:06:51.163847 4953 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Dec 03 00:06:51 crc kubenswrapper[4953]: W1203 00:06:51.163852 4953 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Dec 03 00:06:51 crc kubenswrapper[4953]: W1203 00:06:51.163858 4953 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Dec 03 00:06:51 crc kubenswrapper[4953]: W1203 00:06:51.163873 4953 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Dec 03 00:06:51 crc kubenswrapper[4953]: W1203 00:06:51.163878 4953 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Dec 03 00:06:51 crc kubenswrapper[4953]: W1203 00:06:51.163883 4953 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Dec 03 00:06:51 crc kubenswrapper[4953]: W1203 00:06:51.163890 4953 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Dec 03 00:06:51 crc kubenswrapper[4953]: W1203 00:06:51.163896 4953 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Dec 03 00:06:51 crc kubenswrapper[4953]: W1203 00:06:51.163902 4953 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Dec 03 00:06:51 crc kubenswrapper[4953]: W1203 00:06:51.163907 4953 feature_gate.go:330] unrecognized feature gate: InsightsConfig Dec 03 00:06:51 crc kubenswrapper[4953]: W1203 00:06:51.163912 4953 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Dec 03 00:06:51 crc kubenswrapper[4953]: W1203 00:06:51.163917 4953 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Dec 03 00:06:51 crc kubenswrapper[4953]: W1203 00:06:51.163922 4953 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Dec 03 00:06:51 crc kubenswrapper[4953]: W1203 00:06:51.163927 4953 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Dec 03 00:06:51 crc kubenswrapper[4953]: W1203 00:06:51.163933 4953 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Dec 03 00:06:51 crc kubenswrapper[4953]: W1203 00:06:51.163939 4953 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Dec 03 00:06:51 crc kubenswrapper[4953]: W1203 00:06:51.163945 4953 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Dec 03 00:06:51 crc kubenswrapper[4953]: W1203 00:06:51.163952 4953 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Dec 03 00:06:51 crc kubenswrapper[4953]: W1203 00:06:51.163957 4953 feature_gate.go:330] unrecognized feature gate: PlatformOperators Dec 03 00:06:51 crc kubenswrapper[4953]: W1203 00:06:51.163962 4953 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Dec 03 00:06:51 crc kubenswrapper[4953]: W1203 00:06:51.163968 4953 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Dec 03 00:06:51 crc kubenswrapper[4953]: W1203 00:06:51.163991 4953 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Dec 03 00:06:51 crc kubenswrapper[4953]: W1203 00:06:51.163997 4953 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Dec 03 00:06:51 crc kubenswrapper[4953]: W1203 00:06:51.164003 4953 feature_gate.go:330] unrecognized feature gate: Example Dec 03 00:06:51 crc kubenswrapper[4953]: W1203 00:06:51.164008 4953 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Dec 03 00:06:51 crc kubenswrapper[4953]: W1203 00:06:51.164013 4953 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Dec 03 00:06:51 crc kubenswrapper[4953]: W1203 00:06:51.164018 4953 feature_gate.go:330] unrecognized feature gate: NewOLM Dec 03 00:06:51 crc kubenswrapper[4953]: W1203 00:06:51.164024 4953 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Dec 03 00:06:51 crc kubenswrapper[4953]: W1203 00:06:51.164029 4953 feature_gate.go:330] unrecognized feature gate: GatewayAPI Dec 03 00:06:51 crc kubenswrapper[4953]: W1203 00:06:51.164034 4953 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Dec 03 00:06:51 crc kubenswrapper[4953]: W1203 00:06:51.164039 4953 feature_gate.go:330] unrecognized feature gate: PinnedImages Dec 03 00:06:51 crc kubenswrapper[4953]: W1203 00:06:51.164045 4953 feature_gate.go:330] unrecognized feature gate: OVNObservability Dec 03 00:06:51 crc kubenswrapper[4953]: W1203 00:06:51.164050 4953 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Dec 03 00:06:51 crc kubenswrapper[4953]: W1203 00:06:51.164055 4953 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Dec 03 00:06:51 crc kubenswrapper[4953]: W1203 00:06:51.164061 4953 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Dec 03 00:06:51 crc kubenswrapper[4953]: W1203 00:06:51.164067 4953 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Dec 03 00:06:51 crc kubenswrapper[4953]: W1203 00:06:51.164072 4953 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Dec 03 00:06:51 crc kubenswrapper[4953]: W1203 00:06:51.164078 4953 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Dec 03 00:06:51 crc kubenswrapper[4953]: W1203 00:06:51.164085 4953 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Dec 03 00:06:51 crc kubenswrapper[4953]: W1203 00:06:51.164090 4953 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Dec 03 00:06:51 crc kubenswrapper[4953]: W1203 00:06:51.164095 4953 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Dec 03 00:06:51 crc kubenswrapper[4953]: W1203 00:06:51.164100 4953 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Dec 03 00:06:51 crc kubenswrapper[4953]: W1203 00:06:51.164105 4953 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Dec 03 00:06:51 crc kubenswrapper[4953]: W1203 00:06:51.164110 4953 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Dec 03 00:06:51 crc kubenswrapper[4953]: W1203 00:06:51.164115 4953 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Dec 03 00:06:51 crc kubenswrapper[4953]: W1203 00:06:51.164120 4953 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Dec 03 00:06:51 crc kubenswrapper[4953]: W1203 00:06:51.164125 4953 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Dec 03 00:06:51 crc kubenswrapper[4953]: W1203 00:06:51.164129 4953 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Dec 03 00:06:51 crc kubenswrapper[4953]: W1203 00:06:51.164134 4953 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Dec 03 00:06:51 crc kubenswrapper[4953]: W1203 00:06:51.164162 4953 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Dec 03 00:06:51 crc kubenswrapper[4953]: W1203 00:06:51.164168 4953 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Dec 03 00:06:51 crc kubenswrapper[4953]: W1203 00:06:51.164172 4953 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Dec 03 00:06:51 crc kubenswrapper[4953]: W1203 00:06:51.164177 4953 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Dec 03 00:06:51 crc kubenswrapper[4953]: W1203 00:06:51.164182 4953 feature_gate.go:330] unrecognized feature gate: SignatureStores Dec 03 00:06:51 crc kubenswrapper[4953]: W1203 00:06:51.164187 4953 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Dec 03 00:06:51 crc kubenswrapper[4953]: W1203 00:06:51.164193 4953 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Dec 03 00:06:51 crc kubenswrapper[4953]: W1203 00:06:51.164198 4953 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Dec 03 00:06:51 crc kubenswrapper[4953]: W1203 00:06:51.164203 4953 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Dec 03 00:06:51 crc kubenswrapper[4953]: W1203 00:06:51.164208 4953 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Dec 03 00:06:51 crc kubenswrapper[4953]: W1203 00:06:51.164213 4953 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Dec 03 00:06:51 crc kubenswrapper[4953]: W1203 00:06:51.164217 4953 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.164508 4953 flags.go:64] FLAG: --address="0.0.0.0" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.164521 4953 flags.go:64] FLAG: --allowed-unsafe-sysctls="[]" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.164532 4953 flags.go:64] FLAG: --anonymous-auth="true" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.164540 4953 flags.go:64] FLAG: --application-metrics-count-limit="100" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.164547 4953 flags.go:64] FLAG: --authentication-token-webhook="false" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.164553 4953 flags.go:64] FLAG: --authentication-token-webhook-cache-ttl="2m0s" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.164560 4953 flags.go:64] FLAG: --authorization-mode="AlwaysAllow" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.164567 4953 flags.go:64] FLAG: --authorization-webhook-cache-authorized-ttl="5m0s" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.164574 4953 flags.go:64] FLAG: --authorization-webhook-cache-unauthorized-ttl="30s" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.164580 4953 flags.go:64] FLAG: --boot-id-file="/proc/sys/kernel/random/boot_id" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.164588 4953 flags.go:64] FLAG: --bootstrap-kubeconfig="/etc/kubernetes/kubeconfig" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.164595 4953 flags.go:64] FLAG: --cert-dir="/var/lib/kubelet/pki" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.164602 4953 flags.go:64] FLAG: --cgroup-driver="cgroupfs" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.164608 4953 flags.go:64] FLAG: --cgroup-root="" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.164615 4953 flags.go:64] FLAG: --cgroups-per-qos="true" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.164622 4953 flags.go:64] FLAG: --client-ca-file="" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.164628 4953 flags.go:64] FLAG: --cloud-config="" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.164634 4953 flags.go:64] FLAG: --cloud-provider="" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.164639 4953 flags.go:64] FLAG: --cluster-dns="[]" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.164646 4953 flags.go:64] FLAG: --cluster-domain="" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.164651 4953 flags.go:64] FLAG: --config="/etc/kubernetes/kubelet.conf" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.164657 4953 flags.go:64] FLAG: --config-dir="" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.164662 4953 flags.go:64] FLAG: --container-hints="/etc/cadvisor/container_hints.json" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.164668 4953 flags.go:64] FLAG: --container-log-max-files="5" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.164676 4953 flags.go:64] FLAG: --container-log-max-size="10Mi" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.164681 4953 flags.go:64] FLAG: --container-runtime-endpoint="/var/run/crio/crio.sock" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.164687 4953 flags.go:64] FLAG: --containerd="/run/containerd/containerd.sock" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.164693 4953 flags.go:64] FLAG: --containerd-namespace="k8s.io" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.164698 4953 flags.go:64] FLAG: --contention-profiling="false" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.164704 4953 flags.go:64] FLAG: --cpu-cfs-quota="true" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.164710 4953 flags.go:64] FLAG: --cpu-cfs-quota-period="100ms" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.164716 4953 flags.go:64] FLAG: --cpu-manager-policy="none" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.164722 4953 flags.go:64] FLAG: --cpu-manager-policy-options="" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.164729 4953 flags.go:64] FLAG: --cpu-manager-reconcile-period="10s" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.164735 4953 flags.go:64] FLAG: --enable-controller-attach-detach="true" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.164741 4953 flags.go:64] FLAG: --enable-debugging-handlers="true" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.164747 4953 flags.go:64] FLAG: --enable-load-reader="false" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.164753 4953 flags.go:64] FLAG: --enable-server="true" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.164759 4953 flags.go:64] FLAG: --enforce-node-allocatable="[pods]" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.164767 4953 flags.go:64] FLAG: --event-burst="100" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.164773 4953 flags.go:64] FLAG: --event-qps="50" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.164779 4953 flags.go:64] FLAG: --event-storage-age-limit="default=0" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.164784 4953 flags.go:64] FLAG: --event-storage-event-limit="default=0" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.164790 4953 flags.go:64] FLAG: --eviction-hard="" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.164797 4953 flags.go:64] FLAG: --eviction-max-pod-grace-period="0" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.164803 4953 flags.go:64] FLAG: --eviction-minimum-reclaim="" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.164809 4953 flags.go:64] FLAG: --eviction-pressure-transition-period="5m0s" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.164814 4953 flags.go:64] FLAG: --eviction-soft="" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.164820 4953 flags.go:64] FLAG: --eviction-soft-grace-period="" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.164825 4953 flags.go:64] FLAG: --exit-on-lock-contention="false" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.164831 4953 flags.go:64] FLAG: --experimental-allocatable-ignore-eviction="false" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.164837 4953 flags.go:64] FLAG: --experimental-mounter-path="" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.164842 4953 flags.go:64] FLAG: --fail-cgroupv1="false" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.164847 4953 flags.go:64] FLAG: --fail-swap-on="true" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.164853 4953 flags.go:64] FLAG: --feature-gates="" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.164859 4953 flags.go:64] FLAG: --file-check-frequency="20s" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.164865 4953 flags.go:64] FLAG: --global-housekeeping-interval="1m0s" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.164871 4953 flags.go:64] FLAG: --hairpin-mode="promiscuous-bridge" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.164876 4953 flags.go:64] FLAG: --healthz-bind-address="127.0.0.1" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.164882 4953 flags.go:64] FLAG: --healthz-port="10248" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.164888 4953 flags.go:64] FLAG: --help="false" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.164894 4953 flags.go:64] FLAG: --hostname-override="" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.164899 4953 flags.go:64] FLAG: --housekeeping-interval="10s" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.164905 4953 flags.go:64] FLAG: --http-check-frequency="20s" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.164910 4953 flags.go:64] FLAG: --image-credential-provider-bin-dir="" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.164916 4953 flags.go:64] FLAG: --image-credential-provider-config="" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.164923 4953 flags.go:64] FLAG: --image-gc-high-threshold="85" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.164929 4953 flags.go:64] FLAG: --image-gc-low-threshold="80" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.164935 4953 flags.go:64] FLAG: --image-service-endpoint="" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.164940 4953 flags.go:64] FLAG: --kernel-memcg-notification="false" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.164945 4953 flags.go:64] FLAG: --kube-api-burst="100" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.164951 4953 flags.go:64] FLAG: --kube-api-content-type="application/vnd.kubernetes.protobuf" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.164957 4953 flags.go:64] FLAG: --kube-api-qps="50" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.164962 4953 flags.go:64] FLAG: --kube-reserved="" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.164968 4953 flags.go:64] FLAG: --kube-reserved-cgroup="" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.164995 4953 flags.go:64] FLAG: --kubeconfig="/var/lib/kubelet/kubeconfig" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.165002 4953 flags.go:64] FLAG: --kubelet-cgroups="" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.165007 4953 flags.go:64] FLAG: --local-storage-capacity-isolation="true" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.165013 4953 flags.go:64] FLAG: --lock-file="" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.165018 4953 flags.go:64] FLAG: --log-cadvisor-usage="false" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.165024 4953 flags.go:64] FLAG: --log-flush-frequency="5s" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.165030 4953 flags.go:64] FLAG: --log-json-info-buffer-size="0" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.165038 4953 flags.go:64] FLAG: --log-json-split-stream="false" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.165044 4953 flags.go:64] FLAG: --log-text-info-buffer-size="0" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.165049 4953 flags.go:64] FLAG: --log-text-split-stream="false" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.165055 4953 flags.go:64] FLAG: --logging-format="text" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.165060 4953 flags.go:64] FLAG: --machine-id-file="/etc/machine-id,/var/lib/dbus/machine-id" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.165067 4953 flags.go:64] FLAG: --make-iptables-util-chains="true" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.165072 4953 flags.go:64] FLAG: --manifest-url="" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.165077 4953 flags.go:64] FLAG: --manifest-url-header="" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.165084 4953 flags.go:64] FLAG: --max-housekeeping-interval="15s" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.165090 4953 flags.go:64] FLAG: --max-open-files="1000000" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.165097 4953 flags.go:64] FLAG: --max-pods="110" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.165102 4953 flags.go:64] FLAG: --maximum-dead-containers="-1" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.165108 4953 flags.go:64] FLAG: --maximum-dead-containers-per-container="1" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.165114 4953 flags.go:64] FLAG: --memory-manager-policy="None" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.165119 4953 flags.go:64] FLAG: --minimum-container-ttl-duration="6m0s" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.165125 4953 flags.go:64] FLAG: --minimum-image-ttl-duration="2m0s" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.165132 4953 flags.go:64] FLAG: --node-ip="192.168.126.11" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.165138 4953 flags.go:64] FLAG: --node-labels="node-role.kubernetes.io/control-plane=,node-role.kubernetes.io/master=,node.openshift.io/os_id=rhcos" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.165152 4953 flags.go:64] FLAG: --node-status-max-images="50" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.165158 4953 flags.go:64] FLAG: --node-status-update-frequency="10s" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.165164 4953 flags.go:64] FLAG: --oom-score-adj="-999" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.165170 4953 flags.go:64] FLAG: --pod-cidr="" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.165175 4953 flags.go:64] FLAG: --pod-infra-container-image="quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:33549946e22a9ffa738fd94b1345f90921bc8f92fa6137784cb33c77ad806f9d" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.165184 4953 flags.go:64] FLAG: --pod-manifest-path="" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.165189 4953 flags.go:64] FLAG: --pod-max-pids="-1" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.165195 4953 flags.go:64] FLAG: --pods-per-core="0" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.165201 4953 flags.go:64] FLAG: --port="10250" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.165207 4953 flags.go:64] FLAG: --protect-kernel-defaults="false" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.165212 4953 flags.go:64] FLAG: --provider-id="" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.165217 4953 flags.go:64] FLAG: --qos-reserved="" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.165223 4953 flags.go:64] FLAG: --read-only-port="10255" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.165229 4953 flags.go:64] FLAG: --register-node="true" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.165235 4953 flags.go:64] FLAG: --register-schedulable="true" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.165240 4953 flags.go:64] FLAG: --register-with-taints="node-role.kubernetes.io/master=:NoSchedule" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.165249 4953 flags.go:64] FLAG: --registry-burst="10" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.165255 4953 flags.go:64] FLAG: --registry-qps="5" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.165261 4953 flags.go:64] FLAG: --reserved-cpus="" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.165267 4953 flags.go:64] FLAG: --reserved-memory="" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.165274 4953 flags.go:64] FLAG: --resolv-conf="/etc/resolv.conf" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.165279 4953 flags.go:64] FLAG: --root-dir="/var/lib/kubelet" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.165285 4953 flags.go:64] FLAG: --rotate-certificates="false" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.165291 4953 flags.go:64] FLAG: --rotate-server-certificates="false" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.165297 4953 flags.go:64] FLAG: --runonce="false" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.165303 4953 flags.go:64] FLAG: --runtime-cgroups="/system.slice/crio.service" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.165308 4953 flags.go:64] FLAG: --runtime-request-timeout="2m0s" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.165314 4953 flags.go:64] FLAG: --seccomp-default="false" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.165320 4953 flags.go:64] FLAG: --serialize-image-pulls="true" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.165325 4953 flags.go:64] FLAG: --storage-driver-buffer-duration="1m0s" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.165331 4953 flags.go:64] FLAG: --storage-driver-db="cadvisor" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.165337 4953 flags.go:64] FLAG: --storage-driver-host="localhost:8086" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.165343 4953 flags.go:64] FLAG: --storage-driver-password="root" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.165349 4953 flags.go:64] FLAG: --storage-driver-secure="false" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.165355 4953 flags.go:64] FLAG: --storage-driver-table="stats" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.165360 4953 flags.go:64] FLAG: --storage-driver-user="root" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.165366 4953 flags.go:64] FLAG: --streaming-connection-idle-timeout="4h0m0s" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.165371 4953 flags.go:64] FLAG: --sync-frequency="1m0s" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.165378 4953 flags.go:64] FLAG: --system-cgroups="" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.165383 4953 flags.go:64] FLAG: --system-reserved="cpu=200m,ephemeral-storage=350Mi,memory=350Mi" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.165392 4953 flags.go:64] FLAG: --system-reserved-cgroup="" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.165397 4953 flags.go:64] FLAG: --tls-cert-file="" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.165403 4953 flags.go:64] FLAG: --tls-cipher-suites="[]" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.165409 4953 flags.go:64] FLAG: --tls-min-version="" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.165415 4953 flags.go:64] FLAG: --tls-private-key-file="" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.165420 4953 flags.go:64] FLAG: --topology-manager-policy="none" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.165426 4953 flags.go:64] FLAG: --topology-manager-policy-options="" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.165431 4953 flags.go:64] FLAG: --topology-manager-scope="container" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.165437 4953 flags.go:64] FLAG: --v="2" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.165445 4953 flags.go:64] FLAG: --version="false" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.165452 4953 flags.go:64] FLAG: --vmodule="" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.165459 4953 flags.go:64] FLAG: --volume-plugin-dir="/etc/kubernetes/kubelet-plugins/volume/exec" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.165465 4953 flags.go:64] FLAG: --volume-stats-agg-period="1m0s" Dec 03 00:06:51 crc kubenswrapper[4953]: W1203 00:06:51.165586 4953 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Dec 03 00:06:51 crc kubenswrapper[4953]: W1203 00:06:51.165593 4953 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Dec 03 00:06:51 crc kubenswrapper[4953]: W1203 00:06:51.165598 4953 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Dec 03 00:06:51 crc kubenswrapper[4953]: W1203 00:06:51.165604 4953 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Dec 03 00:06:51 crc kubenswrapper[4953]: W1203 00:06:51.165610 4953 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Dec 03 00:06:51 crc kubenswrapper[4953]: W1203 00:06:51.165615 4953 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Dec 03 00:06:51 crc kubenswrapper[4953]: W1203 00:06:51.165620 4953 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Dec 03 00:06:51 crc kubenswrapper[4953]: W1203 00:06:51.165626 4953 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Dec 03 00:06:51 crc kubenswrapper[4953]: W1203 00:06:51.165630 4953 feature_gate.go:330] unrecognized feature gate: PinnedImages Dec 03 00:06:51 crc kubenswrapper[4953]: W1203 00:06:51.165636 4953 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Dec 03 00:06:51 crc kubenswrapper[4953]: W1203 00:06:51.165646 4953 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Dec 03 00:06:51 crc kubenswrapper[4953]: W1203 00:06:51.165651 4953 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Dec 03 00:06:51 crc kubenswrapper[4953]: W1203 00:06:51.165656 4953 feature_gate.go:330] unrecognized feature gate: NewOLM Dec 03 00:06:51 crc kubenswrapper[4953]: W1203 00:06:51.165661 4953 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Dec 03 00:06:51 crc kubenswrapper[4953]: W1203 00:06:51.165665 4953 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Dec 03 00:06:51 crc kubenswrapper[4953]: W1203 00:06:51.165670 4953 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Dec 03 00:06:51 crc kubenswrapper[4953]: W1203 00:06:51.165675 4953 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Dec 03 00:06:51 crc kubenswrapper[4953]: W1203 00:06:51.165680 4953 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Dec 03 00:06:51 crc kubenswrapper[4953]: W1203 00:06:51.165685 4953 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Dec 03 00:06:51 crc kubenswrapper[4953]: W1203 00:06:51.165690 4953 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Dec 03 00:06:51 crc kubenswrapper[4953]: W1203 00:06:51.165695 4953 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Dec 03 00:06:51 crc kubenswrapper[4953]: W1203 00:06:51.165702 4953 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Dec 03 00:06:51 crc kubenswrapper[4953]: W1203 00:06:51.165708 4953 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Dec 03 00:06:51 crc kubenswrapper[4953]: W1203 00:06:51.165714 4953 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Dec 03 00:06:51 crc kubenswrapper[4953]: W1203 00:06:51.165722 4953 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Dec 03 00:06:51 crc kubenswrapper[4953]: W1203 00:06:51.165729 4953 feature_gate.go:330] unrecognized feature gate: GatewayAPI Dec 03 00:06:51 crc kubenswrapper[4953]: W1203 00:06:51.165736 4953 feature_gate.go:330] unrecognized feature gate: Example Dec 03 00:06:51 crc kubenswrapper[4953]: W1203 00:06:51.165742 4953 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Dec 03 00:06:51 crc kubenswrapper[4953]: W1203 00:06:51.165748 4953 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Dec 03 00:06:51 crc kubenswrapper[4953]: W1203 00:06:51.165753 4953 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Dec 03 00:06:51 crc kubenswrapper[4953]: W1203 00:06:51.165758 4953 feature_gate.go:330] unrecognized feature gate: PlatformOperators Dec 03 00:06:51 crc kubenswrapper[4953]: W1203 00:06:51.165763 4953 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Dec 03 00:06:51 crc kubenswrapper[4953]: W1203 00:06:51.165768 4953 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Dec 03 00:06:51 crc kubenswrapper[4953]: W1203 00:06:51.165773 4953 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Dec 03 00:06:51 crc kubenswrapper[4953]: W1203 00:06:51.165777 4953 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Dec 03 00:06:51 crc kubenswrapper[4953]: W1203 00:06:51.165782 4953 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Dec 03 00:06:51 crc kubenswrapper[4953]: W1203 00:06:51.165787 4953 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Dec 03 00:06:51 crc kubenswrapper[4953]: W1203 00:06:51.165792 4953 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Dec 03 00:06:51 crc kubenswrapper[4953]: W1203 00:06:51.165797 4953 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Dec 03 00:06:51 crc kubenswrapper[4953]: W1203 00:06:51.165801 4953 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Dec 03 00:06:51 crc kubenswrapper[4953]: W1203 00:06:51.165808 4953 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Dec 03 00:06:51 crc kubenswrapper[4953]: W1203 00:06:51.165814 4953 feature_gate.go:330] unrecognized feature gate: OVNObservability Dec 03 00:06:51 crc kubenswrapper[4953]: W1203 00:06:51.165822 4953 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Dec 03 00:06:51 crc kubenswrapper[4953]: W1203 00:06:51.165828 4953 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Dec 03 00:06:51 crc kubenswrapper[4953]: W1203 00:06:51.165833 4953 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Dec 03 00:06:51 crc kubenswrapper[4953]: W1203 00:06:51.165839 4953 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Dec 03 00:06:51 crc kubenswrapper[4953]: W1203 00:06:51.165844 4953 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Dec 03 00:06:51 crc kubenswrapper[4953]: W1203 00:06:51.165849 4953 feature_gate.go:330] unrecognized feature gate: SignatureStores Dec 03 00:06:51 crc kubenswrapper[4953]: W1203 00:06:51.165854 4953 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Dec 03 00:06:51 crc kubenswrapper[4953]: W1203 00:06:51.165859 4953 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Dec 03 00:06:51 crc kubenswrapper[4953]: W1203 00:06:51.165864 4953 feature_gate.go:330] unrecognized feature gate: InsightsConfig Dec 03 00:06:51 crc kubenswrapper[4953]: W1203 00:06:51.165869 4953 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Dec 03 00:06:51 crc kubenswrapper[4953]: W1203 00:06:51.165874 4953 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Dec 03 00:06:51 crc kubenswrapper[4953]: W1203 00:06:51.165878 4953 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Dec 03 00:06:51 crc kubenswrapper[4953]: W1203 00:06:51.165883 4953 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Dec 03 00:06:51 crc kubenswrapper[4953]: W1203 00:06:51.165888 4953 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Dec 03 00:06:51 crc kubenswrapper[4953]: W1203 00:06:51.165893 4953 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Dec 03 00:06:51 crc kubenswrapper[4953]: W1203 00:06:51.165898 4953 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Dec 03 00:06:51 crc kubenswrapper[4953]: W1203 00:06:51.165903 4953 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Dec 03 00:06:51 crc kubenswrapper[4953]: W1203 00:06:51.165908 4953 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Dec 03 00:06:51 crc kubenswrapper[4953]: W1203 00:06:51.165913 4953 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Dec 03 00:06:51 crc kubenswrapper[4953]: W1203 00:06:51.165918 4953 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Dec 03 00:06:51 crc kubenswrapper[4953]: W1203 00:06:51.165922 4953 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Dec 03 00:06:51 crc kubenswrapper[4953]: W1203 00:06:51.165927 4953 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Dec 03 00:06:51 crc kubenswrapper[4953]: W1203 00:06:51.165933 4953 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Dec 03 00:06:51 crc kubenswrapper[4953]: W1203 00:06:51.165937 4953 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Dec 03 00:06:51 crc kubenswrapper[4953]: W1203 00:06:51.165942 4953 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Dec 03 00:06:51 crc kubenswrapper[4953]: W1203 00:06:51.165947 4953 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Dec 03 00:06:51 crc kubenswrapper[4953]: W1203 00:06:51.165952 4953 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Dec 03 00:06:51 crc kubenswrapper[4953]: W1203 00:06:51.165957 4953 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Dec 03 00:06:51 crc kubenswrapper[4953]: W1203 00:06:51.165962 4953 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.165970 4953 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.180683 4953 server.go:491] "Kubelet version" kubeletVersion="v1.31.5" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.180766 4953 server.go:493] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Dec 03 00:06:51 crc kubenswrapper[4953]: W1203 00:06:51.180942 4953 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Dec 03 00:06:51 crc kubenswrapper[4953]: W1203 00:06:51.180963 4953 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Dec 03 00:06:51 crc kubenswrapper[4953]: W1203 00:06:51.181007 4953 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Dec 03 00:06:51 crc kubenswrapper[4953]: W1203 00:06:51.181016 4953 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Dec 03 00:06:51 crc kubenswrapper[4953]: W1203 00:06:51.181025 4953 feature_gate.go:330] unrecognized feature gate: GatewayAPI Dec 03 00:06:51 crc kubenswrapper[4953]: W1203 00:06:51.181033 4953 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Dec 03 00:06:51 crc kubenswrapper[4953]: W1203 00:06:51.181040 4953 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Dec 03 00:06:51 crc kubenswrapper[4953]: W1203 00:06:51.181049 4953 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Dec 03 00:06:51 crc kubenswrapper[4953]: W1203 00:06:51.181056 4953 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Dec 03 00:06:51 crc kubenswrapper[4953]: W1203 00:06:51.181097 4953 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Dec 03 00:06:51 crc kubenswrapper[4953]: W1203 00:06:51.181108 4953 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Dec 03 00:06:51 crc kubenswrapper[4953]: W1203 00:06:51.181118 4953 feature_gate.go:330] unrecognized feature gate: PinnedImages Dec 03 00:06:51 crc kubenswrapper[4953]: W1203 00:06:51.181129 4953 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Dec 03 00:06:51 crc kubenswrapper[4953]: W1203 00:06:51.181143 4953 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Dec 03 00:06:51 crc kubenswrapper[4953]: W1203 00:06:51.181194 4953 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Dec 03 00:06:51 crc kubenswrapper[4953]: W1203 00:06:51.181209 4953 feature_gate.go:330] unrecognized feature gate: NewOLM Dec 03 00:06:51 crc kubenswrapper[4953]: W1203 00:06:51.181219 4953 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Dec 03 00:06:51 crc kubenswrapper[4953]: W1203 00:06:51.181227 4953 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Dec 03 00:06:51 crc kubenswrapper[4953]: W1203 00:06:51.181236 4953 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Dec 03 00:06:51 crc kubenswrapper[4953]: W1203 00:06:51.181244 4953 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Dec 03 00:06:51 crc kubenswrapper[4953]: W1203 00:06:51.181283 4953 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Dec 03 00:06:51 crc kubenswrapper[4953]: W1203 00:06:51.181291 4953 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Dec 03 00:06:51 crc kubenswrapper[4953]: W1203 00:06:51.181299 4953 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Dec 03 00:06:51 crc kubenswrapper[4953]: W1203 00:06:51.181306 4953 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Dec 03 00:06:51 crc kubenswrapper[4953]: W1203 00:06:51.181315 4953 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Dec 03 00:06:51 crc kubenswrapper[4953]: W1203 00:06:51.181323 4953 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Dec 03 00:06:51 crc kubenswrapper[4953]: W1203 00:06:51.181331 4953 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Dec 03 00:06:51 crc kubenswrapper[4953]: W1203 00:06:51.181341 4953 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Dec 03 00:06:51 crc kubenswrapper[4953]: W1203 00:06:51.181377 4953 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Dec 03 00:06:51 crc kubenswrapper[4953]: W1203 00:06:51.181385 4953 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Dec 03 00:06:51 crc kubenswrapper[4953]: W1203 00:06:51.181393 4953 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Dec 03 00:06:51 crc kubenswrapper[4953]: W1203 00:06:51.181401 4953 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Dec 03 00:06:51 crc kubenswrapper[4953]: W1203 00:06:51.181409 4953 feature_gate.go:330] unrecognized feature gate: SignatureStores Dec 03 00:06:51 crc kubenswrapper[4953]: W1203 00:06:51.181417 4953 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Dec 03 00:06:51 crc kubenswrapper[4953]: W1203 00:06:51.181424 4953 feature_gate.go:330] unrecognized feature gate: InsightsConfig Dec 03 00:06:51 crc kubenswrapper[4953]: W1203 00:06:51.181461 4953 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Dec 03 00:06:51 crc kubenswrapper[4953]: W1203 00:06:51.181469 4953 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Dec 03 00:06:51 crc kubenswrapper[4953]: W1203 00:06:51.181477 4953 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Dec 03 00:06:51 crc kubenswrapper[4953]: W1203 00:06:51.181487 4953 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Dec 03 00:06:51 crc kubenswrapper[4953]: W1203 00:06:51.181497 4953 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Dec 03 00:06:51 crc kubenswrapper[4953]: W1203 00:06:51.181511 4953 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Dec 03 00:06:51 crc kubenswrapper[4953]: W1203 00:06:51.181564 4953 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Dec 03 00:06:51 crc kubenswrapper[4953]: W1203 00:06:51.181576 4953 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Dec 03 00:06:51 crc kubenswrapper[4953]: W1203 00:06:51.181586 4953 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Dec 03 00:06:51 crc kubenswrapper[4953]: W1203 00:06:51.181594 4953 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Dec 03 00:06:51 crc kubenswrapper[4953]: W1203 00:06:51.181603 4953 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Dec 03 00:06:51 crc kubenswrapper[4953]: W1203 00:06:51.181611 4953 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Dec 03 00:06:51 crc kubenswrapper[4953]: W1203 00:06:51.181650 4953 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Dec 03 00:06:51 crc kubenswrapper[4953]: W1203 00:06:51.181658 4953 feature_gate.go:330] unrecognized feature gate: Example Dec 03 00:06:51 crc kubenswrapper[4953]: W1203 00:06:51.181666 4953 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Dec 03 00:06:51 crc kubenswrapper[4953]: W1203 00:06:51.181674 4953 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Dec 03 00:06:51 crc kubenswrapper[4953]: W1203 00:06:51.181683 4953 feature_gate.go:330] unrecognized feature gate: PlatformOperators Dec 03 00:06:51 crc kubenswrapper[4953]: W1203 00:06:51.181691 4953 feature_gate.go:330] unrecognized feature gate: OVNObservability Dec 03 00:06:51 crc kubenswrapper[4953]: W1203 00:06:51.181698 4953 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Dec 03 00:06:51 crc kubenswrapper[4953]: W1203 00:06:51.181707 4953 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Dec 03 00:06:51 crc kubenswrapper[4953]: W1203 00:06:51.181745 4953 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Dec 03 00:06:51 crc kubenswrapper[4953]: W1203 00:06:51.181755 4953 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Dec 03 00:06:51 crc kubenswrapper[4953]: W1203 00:06:51.181766 4953 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Dec 03 00:06:51 crc kubenswrapper[4953]: W1203 00:06:51.181777 4953 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Dec 03 00:06:51 crc kubenswrapper[4953]: W1203 00:06:51.181785 4953 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Dec 03 00:06:51 crc kubenswrapper[4953]: W1203 00:06:51.181793 4953 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Dec 03 00:06:51 crc kubenswrapper[4953]: W1203 00:06:51.181828 4953 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Dec 03 00:06:51 crc kubenswrapper[4953]: W1203 00:06:51.181836 4953 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Dec 03 00:06:51 crc kubenswrapper[4953]: W1203 00:06:51.181844 4953 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Dec 03 00:06:51 crc kubenswrapper[4953]: W1203 00:06:51.181853 4953 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Dec 03 00:06:51 crc kubenswrapper[4953]: W1203 00:06:51.181861 4953 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Dec 03 00:06:51 crc kubenswrapper[4953]: W1203 00:06:51.181869 4953 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Dec 03 00:06:51 crc kubenswrapper[4953]: W1203 00:06:51.181877 4953 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Dec 03 00:06:51 crc kubenswrapper[4953]: W1203 00:06:51.181884 4953 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Dec 03 00:06:51 crc kubenswrapper[4953]: W1203 00:06:51.181915 4953 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Dec 03 00:06:51 crc kubenswrapper[4953]: W1203 00:06:51.181924 4953 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.181937 4953 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Dec 03 00:06:51 crc kubenswrapper[4953]: W1203 00:06:51.182312 4953 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Dec 03 00:06:51 crc kubenswrapper[4953]: W1203 00:06:51.182327 4953 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Dec 03 00:06:51 crc kubenswrapper[4953]: W1203 00:06:51.182338 4953 feature_gate.go:330] unrecognized feature gate: SignatureStores Dec 03 00:06:51 crc kubenswrapper[4953]: W1203 00:06:51.182347 4953 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Dec 03 00:06:51 crc kubenswrapper[4953]: W1203 00:06:51.182355 4953 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Dec 03 00:06:51 crc kubenswrapper[4953]: W1203 00:06:51.182363 4953 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Dec 03 00:06:51 crc kubenswrapper[4953]: W1203 00:06:51.182372 4953 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Dec 03 00:06:51 crc kubenswrapper[4953]: W1203 00:06:51.182381 4953 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Dec 03 00:06:51 crc kubenswrapper[4953]: W1203 00:06:51.182388 4953 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Dec 03 00:06:51 crc kubenswrapper[4953]: W1203 00:06:51.182398 4953 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Dec 03 00:06:51 crc kubenswrapper[4953]: W1203 00:06:51.182408 4953 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Dec 03 00:06:51 crc kubenswrapper[4953]: W1203 00:06:51.182448 4953 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Dec 03 00:06:51 crc kubenswrapper[4953]: W1203 00:06:51.182456 4953 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Dec 03 00:06:51 crc kubenswrapper[4953]: W1203 00:06:51.182464 4953 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Dec 03 00:06:51 crc kubenswrapper[4953]: W1203 00:06:51.182471 4953 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Dec 03 00:06:51 crc kubenswrapper[4953]: W1203 00:06:51.182479 4953 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Dec 03 00:06:51 crc kubenswrapper[4953]: W1203 00:06:51.182487 4953 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Dec 03 00:06:51 crc kubenswrapper[4953]: W1203 00:06:51.182494 4953 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Dec 03 00:06:51 crc kubenswrapper[4953]: W1203 00:06:51.182503 4953 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Dec 03 00:06:51 crc kubenswrapper[4953]: W1203 00:06:51.182510 4953 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Dec 03 00:06:51 crc kubenswrapper[4953]: W1203 00:06:51.182521 4953 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Dec 03 00:06:51 crc kubenswrapper[4953]: W1203 00:06:51.182531 4953 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Dec 03 00:06:51 crc kubenswrapper[4953]: W1203 00:06:51.182540 4953 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Dec 03 00:06:51 crc kubenswrapper[4953]: W1203 00:06:51.182549 4953 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Dec 03 00:06:51 crc kubenswrapper[4953]: W1203 00:06:51.182558 4953 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Dec 03 00:06:51 crc kubenswrapper[4953]: W1203 00:06:51.182566 4953 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Dec 03 00:06:51 crc kubenswrapper[4953]: W1203 00:06:51.182574 4953 feature_gate.go:330] unrecognized feature gate: PlatformOperators Dec 03 00:06:51 crc kubenswrapper[4953]: W1203 00:06:51.182584 4953 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Dec 03 00:06:51 crc kubenswrapper[4953]: W1203 00:06:51.182595 4953 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Dec 03 00:06:51 crc kubenswrapper[4953]: W1203 00:06:51.182604 4953 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Dec 03 00:06:51 crc kubenswrapper[4953]: W1203 00:06:51.182613 4953 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Dec 03 00:06:51 crc kubenswrapper[4953]: W1203 00:06:51.182621 4953 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Dec 03 00:06:51 crc kubenswrapper[4953]: W1203 00:06:51.182629 4953 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Dec 03 00:06:51 crc kubenswrapper[4953]: W1203 00:06:51.182638 4953 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Dec 03 00:06:51 crc kubenswrapper[4953]: W1203 00:06:51.182678 4953 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Dec 03 00:06:51 crc kubenswrapper[4953]: W1203 00:06:51.182688 4953 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Dec 03 00:06:51 crc kubenswrapper[4953]: W1203 00:06:51.182699 4953 feature_gate.go:330] unrecognized feature gate: PinnedImages Dec 03 00:06:51 crc kubenswrapper[4953]: W1203 00:06:51.182709 4953 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Dec 03 00:06:51 crc kubenswrapper[4953]: W1203 00:06:51.182722 4953 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Dec 03 00:06:51 crc kubenswrapper[4953]: W1203 00:06:51.182735 4953 feature_gate.go:330] unrecognized feature gate: GatewayAPI Dec 03 00:06:51 crc kubenswrapper[4953]: W1203 00:06:51.182745 4953 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Dec 03 00:06:51 crc kubenswrapper[4953]: W1203 00:06:51.182756 4953 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Dec 03 00:06:51 crc kubenswrapper[4953]: W1203 00:06:51.182768 4953 feature_gate.go:330] unrecognized feature gate: OVNObservability Dec 03 00:06:51 crc kubenswrapper[4953]: W1203 00:06:51.182778 4953 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Dec 03 00:06:51 crc kubenswrapper[4953]: W1203 00:06:51.182789 4953 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Dec 03 00:06:51 crc kubenswrapper[4953]: W1203 00:06:51.182798 4953 feature_gate.go:330] unrecognized feature gate: Example Dec 03 00:06:51 crc kubenswrapper[4953]: W1203 00:06:51.182807 4953 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Dec 03 00:06:51 crc kubenswrapper[4953]: W1203 00:06:51.182817 4953 feature_gate.go:330] unrecognized feature gate: NewOLM Dec 03 00:06:51 crc kubenswrapper[4953]: W1203 00:06:51.182826 4953 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Dec 03 00:06:51 crc kubenswrapper[4953]: W1203 00:06:51.182836 4953 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Dec 03 00:06:51 crc kubenswrapper[4953]: W1203 00:06:51.182846 4953 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Dec 03 00:06:51 crc kubenswrapper[4953]: W1203 00:06:51.182856 4953 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Dec 03 00:06:51 crc kubenswrapper[4953]: W1203 00:06:51.182865 4953 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Dec 03 00:06:51 crc kubenswrapper[4953]: W1203 00:06:51.182875 4953 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Dec 03 00:06:51 crc kubenswrapper[4953]: W1203 00:06:51.182885 4953 feature_gate.go:330] unrecognized feature gate: InsightsConfig Dec 03 00:06:51 crc kubenswrapper[4953]: W1203 00:06:51.182895 4953 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Dec 03 00:06:51 crc kubenswrapper[4953]: W1203 00:06:51.182906 4953 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Dec 03 00:06:51 crc kubenswrapper[4953]: W1203 00:06:51.182915 4953 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Dec 03 00:06:51 crc kubenswrapper[4953]: W1203 00:06:51.182923 4953 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Dec 03 00:06:51 crc kubenswrapper[4953]: W1203 00:06:51.182931 4953 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Dec 03 00:06:51 crc kubenswrapper[4953]: W1203 00:06:51.182939 4953 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Dec 03 00:06:51 crc kubenswrapper[4953]: W1203 00:06:51.182947 4953 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Dec 03 00:06:51 crc kubenswrapper[4953]: W1203 00:06:51.182955 4953 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Dec 03 00:06:51 crc kubenswrapper[4953]: W1203 00:06:51.182964 4953 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Dec 03 00:06:51 crc kubenswrapper[4953]: W1203 00:06:51.182999 4953 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Dec 03 00:06:51 crc kubenswrapper[4953]: W1203 00:06:51.183009 4953 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Dec 03 00:06:51 crc kubenswrapper[4953]: W1203 00:06:51.183017 4953 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Dec 03 00:06:51 crc kubenswrapper[4953]: W1203 00:06:51.183025 4953 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Dec 03 00:06:51 crc kubenswrapper[4953]: W1203 00:06:51.183065 4953 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Dec 03 00:06:51 crc kubenswrapper[4953]: W1203 00:06:51.183073 4953 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Dec 03 00:06:51 crc kubenswrapper[4953]: W1203 00:06:51.183082 4953 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.183094 4953 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.183625 4953 server.go:940] "Client rotation is on, will bootstrap in background" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.188570 4953 bootstrap.go:85] "Current kubeconfig file contents are still valid, no bootstrap necessary" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.188704 4953 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.189584 4953 server.go:997] "Starting client certificate rotation" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.189631 4953 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate rotation is enabled Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.189941 4953 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2026-02-24 05:52:08 +0000 UTC, rotation deadline is 2025-12-30 14:46:01.221659651 +0000 UTC Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.190122 4953 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Waiting 662h39m10.031545049s for next certificate rotation Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.197880 4953 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.201048 4953 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.211105 4953 log.go:25] "Validated CRI v1 runtime API" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.233686 4953 log.go:25] "Validated CRI v1 image API" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.235578 4953 server.go:1437] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.238800 4953 fs.go:133] Filesystem UUIDs: map[0b076daa-c26a-46d2-b3a6-72a8dbc6e257:/dev/vda4 2025-12-03-00-02-30-00:/dev/sr0 7B77-95E7:/dev/vda2 de0497b0-db1b-465a-b278-03db02455c71:/dev/vda3] Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.238846 4953 fs.go:134] Filesystem partitions: map[/dev/shm:{mountpoint:/dev/shm major:0 minor:22 fsType:tmpfs blockSize:0} /dev/vda3:{mountpoint:/boot major:252 minor:3 fsType:ext4 blockSize:0} /dev/vda4:{mountpoint:/var major:252 minor:4 fsType:xfs blockSize:0} /run:{mountpoint:/run major:0 minor:24 fsType:tmpfs blockSize:0} /run/user/1000:{mountpoint:/run/user/1000 major:0 minor:42 fsType:tmpfs blockSize:0} /tmp:{mountpoint:/tmp major:0 minor:30 fsType:tmpfs blockSize:0} /var/lib/etcd:{mountpoint:/var/lib/etcd major:0 minor:43 fsType:tmpfs blockSize:0}] Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.261059 4953 manager.go:217] Machine: {Timestamp:2025-12-03 00:06:51.25966247 +0000 UTC m=+0.243943293 CPUVendorID:AuthenticAMD NumCores:12 NumPhysicalCores:1 NumSockets:12 CpuFrequency:2800000 MemoryCapacity:33654128640 SwapCapacity:0 MemoryByType:map[] NVMInfo:{MemoryModeCapacity:0 AppDirectModeCapacity:0 AvgPowerBudget:0} HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] MachineID:21801e6708c44f15b81395eb736a7cec SystemUUID:4c1f3d44-3d3d-425d-993f-471223206a89 BootID:eb4ae5a6-0f6f-49ef-a027-1cdee9d3c536 Filesystems:[{Device:/var/lib/etcd DeviceMajor:0 DeviceMinor:43 Capacity:1073741824 Type:vfs Inodes:4108170 HasInodes:true} {Device:/dev/shm DeviceMajor:0 DeviceMinor:22 Capacity:16827064320 Type:vfs Inodes:4108170 HasInodes:true} {Device:/run DeviceMajor:0 DeviceMinor:24 Capacity:6730825728 Type:vfs Inodes:819200 HasInodes:true} {Device:/dev/vda4 DeviceMajor:252 DeviceMinor:4 Capacity:85292941312 Type:vfs Inodes:41679680 HasInodes:true} {Device:/tmp DeviceMajor:0 DeviceMinor:30 Capacity:16827064320 Type:vfs Inodes:1048576 HasInodes:true} {Device:/dev/vda3 DeviceMajor:252 DeviceMinor:3 Capacity:366869504 Type:vfs Inodes:98304 HasInodes:true} {Device:/run/user/1000 DeviceMajor:0 DeviceMinor:42 Capacity:3365412864 Type:vfs Inodes:821634 HasInodes:true}] DiskMap:map[252:0:{Name:vda Major:252 Minor:0 Size:214748364800 Scheduler:none}] NetworkDevices:[{Name:br-ex MacAddress:fa:16:3e:b2:40:97 Speed:0 Mtu:1500} {Name:br-int MacAddress:d6:39:55:2e:22:71 Speed:0 Mtu:1400} {Name:ens3 MacAddress:fa:16:3e:b2:40:97 Speed:-1 Mtu:1500} {Name:ens7 MacAddress:fa:16:3e:77:a9:b4 Speed:-1 Mtu:1500} {Name:ens7.20 MacAddress:52:54:00:26:ec:27 Speed:-1 Mtu:1496} {Name:ens7.21 MacAddress:52:54:00:58:d5:1a Speed:-1 Mtu:1496} {Name:ens7.22 MacAddress:52:54:00:71:6f:ea Speed:-1 Mtu:1496} {Name:eth10 MacAddress:ce:70:62:a8:63:84 Speed:0 Mtu:1500} {Name:ovn-k8s-mp0 MacAddress:0a:58:0a:d9:00:02 Speed:0 Mtu:1400} {Name:ovs-system MacAddress:9e:bb:57:3f:da:19 Speed:0 Mtu:1500}] Topology:[{Id:0 Memory:33654128640 HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] Cores:[{Id:0 Threads:[0] Caches:[{Id:0 Size:32768 Type:Data Level:1} {Id:0 Size:32768 Type:Instruction Level:1} {Id:0 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:0 Size:16777216 Type:Unified Level:3}] SocketID:0 BookID: DrawerID:} {Id:0 Threads:[1] Caches:[{Id:1 Size:32768 Type:Data Level:1} {Id:1 Size:32768 Type:Instruction Level:1} {Id:1 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:1 Size:16777216 Type:Unified Level:3}] SocketID:1 BookID: DrawerID:} {Id:0 Threads:[10] Caches:[{Id:10 Size:32768 Type:Data Level:1} {Id:10 Size:32768 Type:Instruction Level:1} {Id:10 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:10 Size:16777216 Type:Unified Level:3}] SocketID:10 BookID: DrawerID:} {Id:0 Threads:[11] Caches:[{Id:11 Size:32768 Type:Data Level:1} {Id:11 Size:32768 Type:Instruction Level:1} {Id:11 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:11 Size:16777216 Type:Unified Level:3}] SocketID:11 BookID: DrawerID:} {Id:0 Threads:[2] Caches:[{Id:2 Size:32768 Type:Data Level:1} {Id:2 Size:32768 Type:Instruction Level:1} {Id:2 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:2 Size:16777216 Type:Unified Level:3}] SocketID:2 BookID: DrawerID:} {Id:0 Threads:[3] Caches:[{Id:3 Size:32768 Type:Data Level:1} {Id:3 Size:32768 Type:Instruction Level:1} {Id:3 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:3 Size:16777216 Type:Unified Level:3}] SocketID:3 BookID: DrawerID:} {Id:0 Threads:[4] Caches:[{Id:4 Size:32768 Type:Data Level:1} {Id:4 Size:32768 Type:Instruction Level:1} {Id:4 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:4 Size:16777216 Type:Unified Level:3}] SocketID:4 BookID: DrawerID:} {Id:0 Threads:[5] Caches:[{Id:5 Size:32768 Type:Data Level:1} {Id:5 Size:32768 Type:Instruction Level:1} {Id:5 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:5 Size:16777216 Type:Unified Level:3}] SocketID:5 BookID: DrawerID:} {Id:0 Threads:[6] Caches:[{Id:6 Size:32768 Type:Data Level:1} {Id:6 Size:32768 Type:Instruction Level:1} {Id:6 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:6 Size:16777216 Type:Unified Level:3}] SocketID:6 BookID: DrawerID:} {Id:0 Threads:[7] Caches:[{Id:7 Size:32768 Type:Data Level:1} {Id:7 Size:32768 Type:Instruction Level:1} {Id:7 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:7 Size:16777216 Type:Unified Level:3}] SocketID:7 BookID: DrawerID:} {Id:0 Threads:[8] Caches:[{Id:8 Size:32768 Type:Data Level:1} {Id:8 Size:32768 Type:Instruction Level:1} {Id:8 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:8 Size:16777216 Type:Unified Level:3}] SocketID:8 BookID: DrawerID:} {Id:0 Threads:[9] Caches:[{Id:9 Size:32768 Type:Data Level:1} {Id:9 Size:32768 Type:Instruction Level:1} {Id:9 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:9 Size:16777216 Type:Unified Level:3}] SocketID:9 BookID: DrawerID:}] Caches:[] Distances:[10]}] CloudProvider:Unknown InstanceType:Unknown InstanceID:None} Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.262311 4953 manager_no_libpfm.go:29] cAdvisor is build without cgo and/or libpfm support. Perf event counters are not available. Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.262714 4953 manager.go:233] Version: {KernelVersion:5.14.0-427.50.2.el9_4.x86_64 ContainerOsVersion:Red Hat Enterprise Linux CoreOS 418.94.202502100215-0 DockerVersion: DockerAPIVersion: CadvisorVersion: CadvisorRevision:} Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.263442 4953 swap_util.go:113] "Swap is on" /proc/swaps contents="Filename\t\t\t\tType\t\tSize\t\tUsed\t\tPriority" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.264120 4953 container_manager_linux.go:267] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.264165 4953 container_manager_linux.go:272] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"crc","RuntimeCgroupsName":"/system.slice/crio.service","SystemCgroupsName":"/system.slice","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":true,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":{"cpu":"200m","ephemeral-storage":"350Mi","memory":"350Mi"},"HardEvictionThresholds":[{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":4096,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.264442 4953 topology_manager.go:138] "Creating topology manager with none policy" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.264455 4953 container_manager_linux.go:303] "Creating device plugin manager" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.264718 4953 manager.go:142] "Creating Device Plugin manager" path="/var/lib/kubelet/device-plugins/kubelet.sock" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.264763 4953 server.go:66] "Creating device plugin registration server" version="v1beta1" socket="/var/lib/kubelet/device-plugins/kubelet.sock" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.265166 4953 state_mem.go:36] "Initialized new in-memory state store" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.265268 4953 server.go:1245] "Using root directory" path="/var/lib/kubelet" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.265937 4953 kubelet.go:418] "Attempting to sync node with API server" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.265958 4953 kubelet.go:313] "Adding static pod path" path="/etc/kubernetes/manifests" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.266004 4953 file.go:69] "Watching path" path="/etc/kubernetes/manifests" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.266021 4953 kubelet.go:324] "Adding apiserver pod source" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.266035 4953 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.268101 4953 kuberuntime_manager.go:262] "Container runtime initialized" containerRuntime="cri-o" version="1.31.5-4.rhaos4.18.gitdad78d5.el9" apiVersion="v1" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.268648 4953 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-server-current.pem". Dec 03 00:06:51 crc kubenswrapper[4953]: W1203 00:06:51.269725 4953 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.158:6443: connect: connection refused Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.269837 4953 kubelet.go:854] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Dec 03 00:06:51 crc kubenswrapper[4953]: E1203 00:06:51.269886 4953 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.158:6443: connect: connection refused" logger="UnhandledError" Dec 03 00:06:51 crc kubenswrapper[4953]: W1203 00:06:51.269732 4953 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.158:6443: connect: connection refused Dec 03 00:06:51 crc kubenswrapper[4953]: E1203 00:06:51.270047 4953 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.158:6443: connect: connection refused" logger="UnhandledError" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.270691 4953 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/portworx-volume" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.270734 4953 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/empty-dir" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.270748 4953 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/git-repo" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.270760 4953 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/host-path" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.270782 4953 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/nfs" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.270795 4953 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/secret" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.270808 4953 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/iscsi" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.270829 4953 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/downward-api" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.270843 4953 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/fc" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.270857 4953 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/configmap" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.270874 4953 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/projected" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.270888 4953 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/local-volume" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.271206 4953 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/csi" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.271860 4953 server.go:1280] "Started kubelet" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.272369 4953 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.158:6443: connect: connection refused Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.272666 4953 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.272795 4953 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.273488 4953 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate rotation is enabled Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.273527 4953 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.273556 4953 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-13 03:38:59.165999812 +0000 UTC Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.273623 4953 certificate_manager.go:356] kubernetes.io/kubelet-serving: Waiting 243h32m7.892385561s for next certificate rotation Dec 03 00:06:51 crc systemd[1]: Started Kubernetes Kubelet. Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.273865 4953 volume_manager.go:287] "The desired_state_of_world populator starts" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.273885 4953 volume_manager.go:289] "Starting Kubelet Volume Manager" Dec 03 00:06:51 crc kubenswrapper[4953]: E1203 00:06:51.273925 4953 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.273965 4953 desired_state_of_world_populator.go:146] "Desired state populator starts to run" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.274383 4953 server.go:236] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.275711 4953 server.go:460] "Adding debug handlers to kubelet server" Dec 03 00:06:51 crc kubenswrapper[4953]: W1203 00:06:51.276927 4953 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.158:6443: connect: connection refused Dec 03 00:06:51 crc kubenswrapper[4953]: E1203 00:06:51.277017 4953 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.158:6443: connect: connection refused" interval="200ms" Dec 03 00:06:51 crc kubenswrapper[4953]: E1203 00:06:51.277049 4953 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.158:6443: connect: connection refused" logger="UnhandledError" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.277391 4953 factory.go:55] Registering systemd factory Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.277417 4953 factory.go:221] Registration of the systemd container factory successfully Dec 03 00:06:51 crc kubenswrapper[4953]: E1203 00:06:51.277225 4953 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": dial tcp 38.102.83.158:6443: connect: connection refused" event="&Event{ObjectMeta:{crc.187d8bdd39bc13ed default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-03 00:06:51.271762925 +0000 UTC m=+0.256043738,LastTimestamp:2025-12-03 00:06:51.271762925 +0000 UTC m=+0.256043738,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.277836 4953 factory.go:153] Registering CRI-O factory Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.277875 4953 factory.go:221] Registration of the crio container factory successfully Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.278024 4953 factory.go:219] Registration of the containerd container factory failed: unable to create containerd client: containerd: cannot unix dial containerd api service: dial unix /run/containerd/containerd.sock: connect: no such file or directory Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.278069 4953 factory.go:103] Registering Raw factory Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.278105 4953 manager.go:1196] Started watching for new ooms in manager Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.282848 4953 manager.go:319] Starting recovery of all containers Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.295306 4953 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" seLinuxMountContext="" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.295382 4953 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" seLinuxMountContext="" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.295404 4953 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" seLinuxMountContext="" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.295426 4953 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" seLinuxMountContext="" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.295445 4953 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" seLinuxMountContext="" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.295463 4953 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" seLinuxMountContext="" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.295481 4953 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" seLinuxMountContext="" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.295499 4953 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" seLinuxMountContext="" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.295522 4953 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" seLinuxMountContext="" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.295540 4953 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" seLinuxMountContext="" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.295567 4953 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" seLinuxMountContext="" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.295586 4953 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" seLinuxMountContext="" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.295608 4953 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" seLinuxMountContext="" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.295635 4953 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" seLinuxMountContext="" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.295653 4953 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" seLinuxMountContext="" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.295669 4953 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" volumeName="kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" seLinuxMountContext="" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.295688 4953 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" seLinuxMountContext="" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.295705 4953 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" seLinuxMountContext="" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.295722 4953 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" seLinuxMountContext="" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.295740 4953 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" seLinuxMountContext="" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.295758 4953 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" seLinuxMountContext="" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.295776 4953 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" seLinuxMountContext="" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.295793 4953 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb" seLinuxMountContext="" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.295812 4953 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" seLinuxMountContext="" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.295829 4953 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" seLinuxMountContext="" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.295846 4953 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="44663579-783b-4372-86d6-acf235a62d72" volumeName="kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" seLinuxMountContext="" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.296000 4953 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" seLinuxMountContext="" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.296036 4953 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" seLinuxMountContext="" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.296094 4953 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" seLinuxMountContext="" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.296115 4953 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" seLinuxMountContext="" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.296132 4953 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" seLinuxMountContext="" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.296150 4953 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" seLinuxMountContext="" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.296168 4953 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" seLinuxMountContext="" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.296185 4953 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" seLinuxMountContext="" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.296237 4953 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" seLinuxMountContext="" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.296257 4953 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" seLinuxMountContext="" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.296276 4953 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" seLinuxMountContext="" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.296294 4953 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides" seLinuxMountContext="" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.296342 4953 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" seLinuxMountContext="" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.296363 4953 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" seLinuxMountContext="" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.296381 4953 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" seLinuxMountContext="" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.296398 4953 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" seLinuxMountContext="" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.296416 4953 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" seLinuxMountContext="" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.296434 4953 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" seLinuxMountContext="" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.296453 4953 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" seLinuxMountContext="" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.296474 4953 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" seLinuxMountContext="" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.296494 4953 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" seLinuxMountContext="" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.296512 4953 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" seLinuxMountContext="" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.296531 4953 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" seLinuxMountContext="" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.296550 4953 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" seLinuxMountContext="" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.296568 4953 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" seLinuxMountContext="" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.296590 4953 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" seLinuxMountContext="" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.296615 4953 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" seLinuxMountContext="" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.296637 4953 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" seLinuxMountContext="" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.296656 4953 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" seLinuxMountContext="" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.296685 4953 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" seLinuxMountContext="" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.296706 4953 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" seLinuxMountContext="" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.296726 4953 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" seLinuxMountContext="" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.296746 4953 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" seLinuxMountContext="" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.296765 4953 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" seLinuxMountContext="" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.296785 4953 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" seLinuxMountContext="" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.296808 4953 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" seLinuxMountContext="" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.296827 4953 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" seLinuxMountContext="" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.296845 4953 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" seLinuxMountContext="" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.296863 4953 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" seLinuxMountContext="" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.296881 4953 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" volumeName="kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" seLinuxMountContext="" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.296899 4953 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" seLinuxMountContext="" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.296918 4953 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3b6479f0-333b-4a96-9adf-2099afdc2447" volumeName="kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr" seLinuxMountContext="" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.296935 4953 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" seLinuxMountContext="" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.296954 4953 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" seLinuxMountContext="" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.297002 4953 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" seLinuxMountContext="" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.297032 4953 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" seLinuxMountContext="" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.297057 4953 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" seLinuxMountContext="" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.297080 4953 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" seLinuxMountContext="" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.297104 4953 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" seLinuxMountContext="" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.297122 4953 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" seLinuxMountContext="" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.297143 4953 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" seLinuxMountContext="" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.297161 4953 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" seLinuxMountContext="" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.297178 4953 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" seLinuxMountContext="" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.297198 4953 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" seLinuxMountContext="" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.297216 4953 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" seLinuxMountContext="" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.297233 4953 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" seLinuxMountContext="" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.297250 4953 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" seLinuxMountContext="" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.297267 4953 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" seLinuxMountContext="" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.297287 4953 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" seLinuxMountContext="" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.297304 4953 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" seLinuxMountContext="" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.297322 4953 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" seLinuxMountContext="" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.297342 4953 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" seLinuxMountContext="" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.297361 4953 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" seLinuxMountContext="" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.297380 4953 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" seLinuxMountContext="" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.297398 4953 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" seLinuxMountContext="" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.297414 4953 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" seLinuxMountContext="" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.297433 4953 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" seLinuxMountContext="" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.297451 4953 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" seLinuxMountContext="" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.297468 4953 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm" seLinuxMountContext="" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.297487 4953 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" seLinuxMountContext="" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.297504 4953 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" seLinuxMountContext="" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.297522 4953 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" seLinuxMountContext="" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.297541 4953 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" seLinuxMountContext="" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.297559 4953 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" seLinuxMountContext="" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.297577 4953 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" seLinuxMountContext="" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.297596 4953 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" seLinuxMountContext="" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.298763 4953 reconstruct.go:144] "Volume is marked device as uncertain and added into the actual state" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" deviceMountPath="/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.299665 4953 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" seLinuxMountContext="" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.299702 4953 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" seLinuxMountContext="" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.299733 4953 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" seLinuxMountContext="" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.299755 4953 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" seLinuxMountContext="" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.299775 4953 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" seLinuxMountContext="" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.299795 4953 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" seLinuxMountContext="" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.299816 4953 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" seLinuxMountContext="" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.299838 4953 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" seLinuxMountContext="" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.299857 4953 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" seLinuxMountContext="" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.299876 4953 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" seLinuxMountContext="" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.299897 4953 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" seLinuxMountContext="" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.299917 4953 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" seLinuxMountContext="" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.300061 4953 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" seLinuxMountContext="" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.300120 4953 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script" seLinuxMountContext="" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.300167 4953 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" seLinuxMountContext="" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.300185 4953 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" seLinuxMountContext="" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.300271 4953 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" seLinuxMountContext="" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.300291 4953 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" seLinuxMountContext="" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.300335 4953 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" seLinuxMountContext="" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.300380 4953 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" seLinuxMountContext="" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.300399 4953 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" seLinuxMountContext="" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.300446 4953 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" seLinuxMountContext="" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.300499 4953 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" seLinuxMountContext="" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.300521 4953 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" seLinuxMountContext="" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.300542 4953 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5" seLinuxMountContext="" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.300564 4953 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert" seLinuxMountContext="" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.300607 4953 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" seLinuxMountContext="" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.300656 4953 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" seLinuxMountContext="" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.300681 4953 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" seLinuxMountContext="" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.300743 4953 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" seLinuxMountContext="" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.300787 4953 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" seLinuxMountContext="" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.300845 4953 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" seLinuxMountContext="" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.300889 4953 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" seLinuxMountContext="" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.300927 4953 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" seLinuxMountContext="" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.304526 4953 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf" seLinuxMountContext="" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.305149 4953 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" seLinuxMountContext="" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.305187 4953 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" seLinuxMountContext="" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.305236 4953 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" seLinuxMountContext="" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.305272 4953 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" seLinuxMountContext="" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.305300 4953 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" seLinuxMountContext="" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.305330 4953 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" seLinuxMountContext="" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.305360 4953 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" seLinuxMountContext="" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.305390 4953 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" seLinuxMountContext="" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.305420 4953 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" seLinuxMountContext="" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.305450 4953 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" seLinuxMountContext="" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.305478 4953 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" seLinuxMountContext="" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.305510 4953 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" seLinuxMountContext="" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.305540 4953 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" seLinuxMountContext="" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.305571 4953 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" seLinuxMountContext="" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.305601 4953 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" seLinuxMountContext="" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.305701 4953 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" seLinuxMountContext="" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.305734 4953 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49ef4625-1d3a-4a9f-b595-c2433d32326d" volumeName="kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" seLinuxMountContext="" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.305850 4953 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" seLinuxMountContext="" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.305947 4953 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" seLinuxMountContext="" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.306035 4953 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" seLinuxMountContext="" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.306064 4953 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" seLinuxMountContext="" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.306134 4953 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" seLinuxMountContext="" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.306154 4953 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert" seLinuxMountContext="" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.306206 4953 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" seLinuxMountContext="" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.306227 4953 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" seLinuxMountContext="" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.306246 4953 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" seLinuxMountContext="" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.306298 4953 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" seLinuxMountContext="" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.306321 4953 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" seLinuxMountContext="" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.306373 4953 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" seLinuxMountContext="" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.306393 4953 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" seLinuxMountContext="" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.306415 4953 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" seLinuxMountContext="" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.306469 4953 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" seLinuxMountContext="" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.306493 4953 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" seLinuxMountContext="" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.306514 4953 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" seLinuxMountContext="" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.306566 4953 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" seLinuxMountContext="" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.306587 4953 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" seLinuxMountContext="" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.306605 4953 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" seLinuxMountContext="" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.306660 4953 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" seLinuxMountContext="" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.306683 4953 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" seLinuxMountContext="" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.306748 4953 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" seLinuxMountContext="" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.306779 4953 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" seLinuxMountContext="" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.306843 4953 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d751cbb-f2e2-430d-9754-c882a5e924a5" volumeName="kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl" seLinuxMountContext="" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.306871 4953 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" seLinuxMountContext="" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.306932 4953 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" seLinuxMountContext="" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.306960 4953 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" seLinuxMountContext="" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.307038 4953 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" seLinuxMountContext="" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.307068 4953 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" seLinuxMountContext="" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.307142 4953 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf" seLinuxMountContext="" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.307174 4953 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" seLinuxMountContext="" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.307248 4953 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" seLinuxMountContext="" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.307316 4953 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" seLinuxMountContext="" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.307348 4953 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" seLinuxMountContext="" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.307418 4953 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" seLinuxMountContext="" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.307448 4953 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" seLinuxMountContext="" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.307528 4953 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" seLinuxMountContext="" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.307563 4953 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" seLinuxMountContext="" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.307644 4953 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" seLinuxMountContext="" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.307726 4953 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" seLinuxMountContext="" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.307757 4953 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" seLinuxMountContext="" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.307827 4953 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" seLinuxMountContext="" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.307859 4953 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" seLinuxMountContext="" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.307926 4953 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" seLinuxMountContext="" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.307957 4953 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" seLinuxMountContext="" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.308040 4953 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" seLinuxMountContext="" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.308072 4953 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" seLinuxMountContext="" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.308144 4953 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" seLinuxMountContext="" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.308172 4953 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" seLinuxMountContext="" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.308288 4953 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls" seLinuxMountContext="" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.308357 4953 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" seLinuxMountContext="" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.308385 4953 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" seLinuxMountContext="" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.308450 4953 reconstruct.go:97] "Volume reconstruction finished" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.308469 4953 reconciler.go:26] "Reconciler: start to sync state" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.312736 4953 manager.go:324] Recovery completed Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.327623 4953 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.329107 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.329227 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.329295 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.330517 4953 cpu_manager.go:225] "Starting CPU manager" policy="none" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.330543 4953 cpu_manager.go:226] "Reconciling" reconcilePeriod="10s" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.330567 4953 state_mem.go:36] "Initialized new in-memory state store" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.334548 4953 policy_none.go:49] "None policy: Start" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.336020 4953 memory_manager.go:170] "Starting memorymanager" policy="None" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.336061 4953 state_mem.go:35] "Initializing new in-memory state store" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.341998 4953 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.346688 4953 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.346740 4953 status_manager.go:217] "Starting to sync pod status with apiserver" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.346777 4953 kubelet.go:2335] "Starting kubelet main sync loop" Dec 03 00:06:51 crc kubenswrapper[4953]: E1203 00:06:51.346905 4953 kubelet.go:2359] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Dec 03 00:06:51 crc kubenswrapper[4953]: W1203 00:06:51.347963 4953 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.158:6443: connect: connection refused Dec 03 00:06:51 crc kubenswrapper[4953]: E1203 00:06:51.348061 4953 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.158:6443: connect: connection refused" logger="UnhandledError" Dec 03 00:06:51 crc kubenswrapper[4953]: E1203 00:06:51.374053 4953 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.398668 4953 manager.go:334] "Starting Device Plugin manager" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.398742 4953 manager.go:513] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.398764 4953 server.go:79] "Starting device plugin registration server" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.399532 4953 eviction_manager.go:189] "Eviction manager: starting control loop" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.399666 4953 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.399941 4953 plugin_watcher.go:51] "Plugin Watcher Start" path="/var/lib/kubelet/plugins_registry" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.400118 4953 plugin_manager.go:116] "The desired_state_of_world populator (plugin watcher) starts" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.400143 4953 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Dec 03 00:06:51 crc kubenswrapper[4953]: E1203 00:06:51.408609 4953 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.447569 4953 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc","openshift-etcd/etcd-crc","openshift-kube-apiserver/kube-apiserver-crc","openshift-kube-controller-manager/kube-controller-manager-crc","openshift-kube-scheduler/openshift-kube-scheduler-crc"] Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.447652 4953 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.448788 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.448832 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.448866 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.449137 4953 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.449460 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.449518 4953 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.450409 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.450462 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.450479 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.450661 4953 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.450902 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.450933 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.450981 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.450965 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.451072 4953 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.451890 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.451927 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.451962 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.452199 4953 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.452438 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.452499 4953 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.452563 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.452589 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.452600 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.454734 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.454760 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.454771 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.454815 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.454856 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.454878 4953 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.454945 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.454993 4953 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.454881 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.455793 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.455823 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.455827 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.455838 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.455881 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.455915 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.456210 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.456311 4953 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.457495 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.457567 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.457595 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:06:51 crc kubenswrapper[4953]: E1203 00:06:51.478409 4953 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.158:6443: connect: connection refused" interval="400ms" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.500799 4953 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.502398 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.502445 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.502462 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.502498 4953 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 03 00:06:51 crc kubenswrapper[4953]: E1203 00:06:51.503134 4953 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.158:6443: connect: connection refused" node="crc" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.511791 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.511847 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.511890 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.511950 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.512009 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.512031 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.512049 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.512068 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.512087 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.512106 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.512129 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.512181 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.512255 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.512355 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.512397 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.614013 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.614096 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.614135 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.614162 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.614193 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.614228 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.614262 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.614291 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.614324 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.614352 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.614381 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.614409 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.614438 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.614466 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.614494 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.615175 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.615307 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.615346 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.615378 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.615445 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.615498 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.615547 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.615594 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.615640 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.615685 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.615746 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.615790 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.615837 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.615935 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.616004 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.703931 4953 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.706200 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.706414 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.706584 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.706795 4953 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 03 00:06:51 crc kubenswrapper[4953]: E1203 00:06:51.707712 4953 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.158:6443: connect: connection refused" node="crc" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.779424 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.804055 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Dec 03 00:06:51 crc kubenswrapper[4953]: W1203 00:06:51.813939 4953 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd1b160f5dda77d281dd8e69ec8d817f9.slice/crio-90afc1cbd97fe111975e2977984e5718f7ad8a439d91668d9c4061f351ffe394 WatchSource:0}: Error finding container 90afc1cbd97fe111975e2977984e5718f7ad8a439d91668d9c4061f351ffe394: Status 404 returned error can't find the container with id 90afc1cbd97fe111975e2977984e5718f7ad8a439d91668d9c4061f351ffe394 Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.815167 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.844303 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 03 00:06:51 crc kubenswrapper[4953]: I1203 00:06:51.853395 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 03 00:06:51 crc kubenswrapper[4953]: W1203 00:06:51.868255 4953 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf614b9022728cf315e60c057852e563e.slice/crio-e2feb3702771820e8fdcd8775799a00b7f364a1b286eb3f8bfd2cb0fe269a3f6 WatchSource:0}: Error finding container e2feb3702771820e8fdcd8775799a00b7f364a1b286eb3f8bfd2cb0fe269a3f6: Status 404 returned error can't find the container with id e2feb3702771820e8fdcd8775799a00b7f364a1b286eb3f8bfd2cb0fe269a3f6 Dec 03 00:06:51 crc kubenswrapper[4953]: E1203 00:06:51.879207 4953 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.158:6443: connect: connection refused" interval="800ms" Dec 03 00:06:51 crc kubenswrapper[4953]: W1203 00:06:51.889145 4953 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3dcd261975c3d6b9a6ad6367fd4facd3.slice/crio-e006773ea6d198a9e0595c954fa32a22b67589f8cda22307a4a41d17f1ee6d29 WatchSource:0}: Error finding container e006773ea6d198a9e0595c954fa32a22b67589f8cda22307a4a41d17f1ee6d29: Status 404 returned error can't find the container with id e006773ea6d198a9e0595c954fa32a22b67589f8cda22307a4a41d17f1ee6d29 Dec 03 00:06:52 crc kubenswrapper[4953]: I1203 00:06:52.108676 4953 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 00:06:52 crc kubenswrapper[4953]: I1203 00:06:52.110192 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:06:52 crc kubenswrapper[4953]: I1203 00:06:52.110244 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:06:52 crc kubenswrapper[4953]: I1203 00:06:52.110297 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:06:52 crc kubenswrapper[4953]: I1203 00:06:52.110332 4953 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 03 00:06:52 crc kubenswrapper[4953]: E1203 00:06:52.110818 4953 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.158:6443: connect: connection refused" node="crc" Dec 03 00:06:52 crc kubenswrapper[4953]: W1203 00:06:52.273128 4953 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.158:6443: connect: connection refused Dec 03 00:06:52 crc kubenswrapper[4953]: E1203 00:06:52.273199 4953 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.158:6443: connect: connection refused" logger="UnhandledError" Dec 03 00:06:52 crc kubenswrapper[4953]: I1203 00:06:52.273273 4953 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.158:6443: connect: connection refused Dec 03 00:06:52 crc kubenswrapper[4953]: I1203 00:06:52.357001 4953 generic.go:334] "Generic (PLEG): container finished" podID="3dcd261975c3d6b9a6ad6367fd4facd3" containerID="d1d60162f7f6885d4f51b4ab44163cfae6d2d04355b1dd710e26b022450005df" exitCode=0 Dec 03 00:06:52 crc kubenswrapper[4953]: I1203 00:06:52.357102 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerDied","Data":"d1d60162f7f6885d4f51b4ab44163cfae6d2d04355b1dd710e26b022450005df"} Dec 03 00:06:52 crc kubenswrapper[4953]: I1203 00:06:52.357258 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"e006773ea6d198a9e0595c954fa32a22b67589f8cda22307a4a41d17f1ee6d29"} Dec 03 00:06:52 crc kubenswrapper[4953]: I1203 00:06:52.357354 4953 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 00:06:52 crc kubenswrapper[4953]: I1203 00:06:52.358320 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:06:52 crc kubenswrapper[4953]: I1203 00:06:52.358340 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:06:52 crc kubenswrapper[4953]: I1203 00:06:52.358350 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:06:52 crc kubenswrapper[4953]: I1203 00:06:52.358605 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"a2810ffc1b141fa7da811406053db9d72c708edf5f2bb1dd75e4d52ab35fd24c"} Dec 03 00:06:52 crc kubenswrapper[4953]: I1203 00:06:52.358643 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"e2feb3702771820e8fdcd8775799a00b7f364a1b286eb3f8bfd2cb0fe269a3f6"} Dec 03 00:06:52 crc kubenswrapper[4953]: I1203 00:06:52.361875 4953 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="72af6c27296eae2e275d455795c68f0d272869d63642b8353b17d466d0c901cf" exitCode=0 Dec 03 00:06:52 crc kubenswrapper[4953]: I1203 00:06:52.361924 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"72af6c27296eae2e275d455795c68f0d272869d63642b8353b17d466d0c901cf"} Dec 03 00:06:52 crc kubenswrapper[4953]: I1203 00:06:52.361942 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"d50e6232e0908069c9486ff936c3704d7ce64997212e48ff33bee39190219d54"} Dec 03 00:06:52 crc kubenswrapper[4953]: I1203 00:06:52.362031 4953 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 00:06:52 crc kubenswrapper[4953]: I1203 00:06:52.362877 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:06:52 crc kubenswrapper[4953]: I1203 00:06:52.362901 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:06:52 crc kubenswrapper[4953]: I1203 00:06:52.362910 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:06:52 crc kubenswrapper[4953]: I1203 00:06:52.363783 4953 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="e368e6f4b30c99dc160ccf1c7d06b58ad81501ff59c26e7d666fc28f4a097c25" exitCode=0 Dec 03 00:06:52 crc kubenswrapper[4953]: I1203 00:06:52.363833 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"e368e6f4b30c99dc160ccf1c7d06b58ad81501ff59c26e7d666fc28f4a097c25"} Dec 03 00:06:52 crc kubenswrapper[4953]: I1203 00:06:52.363850 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"5c269c4cbcd1e58e4ac839a128e44056549edb19535c655b52db034edd3f9de7"} Dec 03 00:06:52 crc kubenswrapper[4953]: I1203 00:06:52.363929 4953 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 00:06:52 crc kubenswrapper[4953]: I1203 00:06:52.364756 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:06:52 crc kubenswrapper[4953]: I1203 00:06:52.364826 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:06:52 crc kubenswrapper[4953]: I1203 00:06:52.364844 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:06:52 crc kubenswrapper[4953]: I1203 00:06:52.366316 4953 generic.go:334] "Generic (PLEG): container finished" podID="d1b160f5dda77d281dd8e69ec8d817f9" containerID="2ad2a6f5a3296c6a4f53f9da696e3e3f4d6c919f44b380c9b4114021b8ad74b6" exitCode=0 Dec 03 00:06:52 crc kubenswrapper[4953]: I1203 00:06:52.366620 4953 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 00:06:52 crc kubenswrapper[4953]: I1203 00:06:52.366924 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerDied","Data":"2ad2a6f5a3296c6a4f53f9da696e3e3f4d6c919f44b380c9b4114021b8ad74b6"} Dec 03 00:06:52 crc kubenswrapper[4953]: I1203 00:06:52.366964 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"90afc1cbd97fe111975e2977984e5718f7ad8a439d91668d9c4061f351ffe394"} Dec 03 00:06:52 crc kubenswrapper[4953]: I1203 00:06:52.367061 4953 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 00:06:52 crc kubenswrapper[4953]: I1203 00:06:52.367793 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:06:52 crc kubenswrapper[4953]: I1203 00:06:52.367823 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:06:52 crc kubenswrapper[4953]: I1203 00:06:52.367837 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:06:52 crc kubenswrapper[4953]: I1203 00:06:52.368230 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:06:52 crc kubenswrapper[4953]: I1203 00:06:52.368261 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:06:52 crc kubenswrapper[4953]: I1203 00:06:52.368278 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:06:52 crc kubenswrapper[4953]: W1203 00:06:52.514309 4953 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.158:6443: connect: connection refused Dec 03 00:06:52 crc kubenswrapper[4953]: E1203 00:06:52.514386 4953 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.158:6443: connect: connection refused" logger="UnhandledError" Dec 03 00:06:52 crc kubenswrapper[4953]: E1203 00:06:52.680267 4953 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.158:6443: connect: connection refused" interval="1.6s" Dec 03 00:06:52 crc kubenswrapper[4953]: W1203 00:06:52.724941 4953 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.158:6443: connect: connection refused Dec 03 00:06:52 crc kubenswrapper[4953]: E1203 00:06:52.725025 4953 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.158:6443: connect: connection refused" logger="UnhandledError" Dec 03 00:06:52 crc kubenswrapper[4953]: W1203 00:06:52.784525 4953 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.158:6443: connect: connection refused Dec 03 00:06:52 crc kubenswrapper[4953]: E1203 00:06:52.784581 4953 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.158:6443: connect: connection refused" logger="UnhandledError" Dec 03 00:06:52 crc kubenswrapper[4953]: I1203 00:06:52.912770 4953 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 00:06:52 crc kubenswrapper[4953]: I1203 00:06:52.916105 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:06:52 crc kubenswrapper[4953]: I1203 00:06:52.916134 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:06:52 crc kubenswrapper[4953]: I1203 00:06:52.916142 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:06:52 crc kubenswrapper[4953]: I1203 00:06:52.916162 4953 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 03 00:06:52 crc kubenswrapper[4953]: E1203 00:06:52.916496 4953 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.158:6443: connect: connection refused" node="crc" Dec 03 00:06:53 crc kubenswrapper[4953]: I1203 00:06:53.373477 4953 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="1e7ac4753c90c2bece705c1a10cdba5dd02474b35c635f0395b24cf05b76e71a" exitCode=0 Dec 03 00:06:53 crc kubenswrapper[4953]: I1203 00:06:53.373604 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"1e7ac4753c90c2bece705c1a10cdba5dd02474b35c635f0395b24cf05b76e71a"} Dec 03 00:06:53 crc kubenswrapper[4953]: I1203 00:06:53.373820 4953 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 00:06:53 crc kubenswrapper[4953]: I1203 00:06:53.375030 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:06:53 crc kubenswrapper[4953]: I1203 00:06:53.375064 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:06:53 crc kubenswrapper[4953]: I1203 00:06:53.375074 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:06:53 crc kubenswrapper[4953]: I1203 00:06:53.377489 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"e4c95eb1f9b0c797376a1b557415b3f956890c971dc77c5257f238b4f32ff9b5"} Dec 03 00:06:53 crc kubenswrapper[4953]: I1203 00:06:53.377621 4953 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 00:06:53 crc kubenswrapper[4953]: I1203 00:06:53.378801 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:06:53 crc kubenswrapper[4953]: I1203 00:06:53.378861 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:06:53 crc kubenswrapper[4953]: I1203 00:06:53.378884 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:06:53 crc kubenswrapper[4953]: I1203 00:06:53.380256 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"650b6c79cda2e1768ba1d8763db799e210ce2748893512068d25e60f821937ed"} Dec 03 00:06:53 crc kubenswrapper[4953]: I1203 00:06:53.380307 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"5ab6052f1abf17225155ce1bd86afa742e6e78f507b7deedd7161cd8fba9120b"} Dec 03 00:06:53 crc kubenswrapper[4953]: I1203 00:06:53.380327 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"2030aef5b7170991268fa5dc7e7ed8ce1fbe3a2765165a10c25251426575330f"} Dec 03 00:06:53 crc kubenswrapper[4953]: I1203 00:06:53.380349 4953 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 00:06:53 crc kubenswrapper[4953]: I1203 00:06:53.381696 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:06:53 crc kubenswrapper[4953]: I1203 00:06:53.381722 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:06:53 crc kubenswrapper[4953]: I1203 00:06:53.381734 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:06:53 crc kubenswrapper[4953]: I1203 00:06:53.383261 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"258d78e1ad9d3dcda6976909ae3fe71aecead5b995fc66b3fd6e2722d80b5ab8"} Dec 03 00:06:53 crc kubenswrapper[4953]: I1203 00:06:53.383355 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"c67b4277b20a9f1287f50ef82d0785c9d02543a8e7ed26e0413eae332b866db7"} Dec 03 00:06:53 crc kubenswrapper[4953]: I1203 00:06:53.383381 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"72c5cd7dc1c3a8adfedab6ba43a59d80b3efe65f2bb56be95e76cc86a409dcf3"} Dec 03 00:06:53 crc kubenswrapper[4953]: I1203 00:06:53.383522 4953 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 00:06:53 crc kubenswrapper[4953]: I1203 00:06:53.384618 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:06:53 crc kubenswrapper[4953]: I1203 00:06:53.384653 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:06:53 crc kubenswrapper[4953]: I1203 00:06:53.384669 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:06:53 crc kubenswrapper[4953]: I1203 00:06:53.388382 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"761973393a93c59c23e69f41b8ac0d16abc35abd61ecbc6d51a42d94cd664233"} Dec 03 00:06:53 crc kubenswrapper[4953]: I1203 00:06:53.388422 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"3c143d4232fb8bf7ef5999a664e431a00ab0db030f4d31a28d0bb21e00e38581"} Dec 03 00:06:53 crc kubenswrapper[4953]: I1203 00:06:53.388442 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"29e4b9d1b70a93fb133e0759ad148ae133766a66efeae3ff06a21dfae0e805ff"} Dec 03 00:06:53 crc kubenswrapper[4953]: I1203 00:06:53.388460 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"dec43536772aa7c3f5a662fa7f3bd73b1cae2791ad454f85792287828d0ffeed"} Dec 03 00:06:53 crc kubenswrapper[4953]: I1203 00:06:53.388478 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"cae5e2994fcc24dc9b2ea73e79572615b3311edc42642b7f278fb84123989b23"} Dec 03 00:06:53 crc kubenswrapper[4953]: I1203 00:06:53.388594 4953 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 00:06:53 crc kubenswrapper[4953]: I1203 00:06:53.389440 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:06:53 crc kubenswrapper[4953]: I1203 00:06:53.389474 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:06:53 crc kubenswrapper[4953]: I1203 00:06:53.389490 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:06:53 crc kubenswrapper[4953]: I1203 00:06:53.807776 4953 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 00:06:54 crc kubenswrapper[4953]: I1203 00:06:54.395480 4953 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="02c732b24912abc09cdbc2e1a73246e9aea159953958fd808ec30a1b0af1ab95" exitCode=0 Dec 03 00:06:54 crc kubenswrapper[4953]: I1203 00:06:54.395671 4953 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 00:06:54 crc kubenswrapper[4953]: I1203 00:06:54.396087 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"02c732b24912abc09cdbc2e1a73246e9aea159953958fd808ec30a1b0af1ab95"} Dec 03 00:06:54 crc kubenswrapper[4953]: I1203 00:06:54.396265 4953 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 00:06:54 crc kubenswrapper[4953]: I1203 00:06:54.396273 4953 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 00:06:54 crc kubenswrapper[4953]: I1203 00:06:54.396370 4953 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 00:06:54 crc kubenswrapper[4953]: I1203 00:06:54.396570 4953 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 03 00:06:54 crc kubenswrapper[4953]: I1203 00:06:54.398205 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:06:54 crc kubenswrapper[4953]: I1203 00:06:54.398232 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:06:54 crc kubenswrapper[4953]: I1203 00:06:54.398324 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:06:54 crc kubenswrapper[4953]: I1203 00:06:54.398353 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:06:54 crc kubenswrapper[4953]: I1203 00:06:54.398380 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:06:54 crc kubenswrapper[4953]: I1203 00:06:54.398279 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:06:54 crc kubenswrapper[4953]: I1203 00:06:54.398418 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:06:54 crc kubenswrapper[4953]: I1203 00:06:54.398446 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:06:54 crc kubenswrapper[4953]: I1203 00:06:54.398424 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:06:54 crc kubenswrapper[4953]: I1203 00:06:54.398933 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:06:54 crc kubenswrapper[4953]: I1203 00:06:54.398949 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:06:54 crc kubenswrapper[4953]: I1203 00:06:54.398958 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:06:54 crc kubenswrapper[4953]: I1203 00:06:54.436641 4953 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 00:06:54 crc kubenswrapper[4953]: I1203 00:06:54.517330 4953 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 00:06:54 crc kubenswrapper[4953]: I1203 00:06:54.518834 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:06:54 crc kubenswrapper[4953]: I1203 00:06:54.518906 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:06:54 crc kubenswrapper[4953]: I1203 00:06:54.518925 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:06:54 crc kubenswrapper[4953]: I1203 00:06:54.518966 4953 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 03 00:06:55 crc kubenswrapper[4953]: I1203 00:06:55.164106 4953 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 03 00:06:55 crc kubenswrapper[4953]: I1203 00:06:55.401734 4953 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 00:06:55 crc kubenswrapper[4953]: I1203 00:06:55.401812 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"0f89be761666af597188bc8eee98d79e309f9a6ebd78911acd93bb64117ce58f"} Dec 03 00:06:55 crc kubenswrapper[4953]: I1203 00:06:55.401883 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"cd4eaa5e7f308492271c081733e8024ed742600a8b382afaca7b44bd4b34eee4"} Dec 03 00:06:55 crc kubenswrapper[4953]: I1203 00:06:55.401908 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"7cb93b828d224803aa0ffd48c8db717c0634da30bc979833522152fbaab1688d"} Dec 03 00:06:55 crc kubenswrapper[4953]: I1203 00:06:55.401925 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"ca5bca7ce3aecf7d1b1c39ec764575c91a1e69bc9f25184307e6c86584c8893d"} Dec 03 00:06:55 crc kubenswrapper[4953]: I1203 00:06:55.401842 4953 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 00:06:55 crc kubenswrapper[4953]: I1203 00:06:55.402765 4953 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 00:06:55 crc kubenswrapper[4953]: I1203 00:06:55.402871 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:06:55 crc kubenswrapper[4953]: I1203 00:06:55.402903 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:06:55 crc kubenswrapper[4953]: I1203 00:06:55.402914 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:06:55 crc kubenswrapper[4953]: I1203 00:06:55.403681 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:06:55 crc kubenswrapper[4953]: I1203 00:06:55.403704 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:06:55 crc kubenswrapper[4953]: I1203 00:06:55.403713 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:06:55 crc kubenswrapper[4953]: I1203 00:06:55.404103 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:06:55 crc kubenswrapper[4953]: I1203 00:06:55.404169 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:06:55 crc kubenswrapper[4953]: I1203 00:06:55.404189 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:06:56 crc kubenswrapper[4953]: I1203 00:06:56.411683 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"6fa6a7a9a7a52d44817874ca6dd47b8a60306ec313d4a18de8114703e239b71e"} Dec 03 00:06:56 crc kubenswrapper[4953]: I1203 00:06:56.411914 4953 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 00:06:56 crc kubenswrapper[4953]: I1203 00:06:56.413286 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:06:56 crc kubenswrapper[4953]: I1203 00:06:56.413336 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:06:56 crc kubenswrapper[4953]: I1203 00:06:56.413354 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:06:56 crc kubenswrapper[4953]: I1203 00:06:56.862761 4953 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 03 00:06:56 crc kubenswrapper[4953]: I1203 00:06:56.863307 4953 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 00:06:56 crc kubenswrapper[4953]: I1203 00:06:56.865782 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:06:56 crc kubenswrapper[4953]: I1203 00:06:56.865856 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:06:56 crc kubenswrapper[4953]: I1203 00:06:56.865882 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:06:57 crc kubenswrapper[4953]: I1203 00:06:57.275924 4953 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-etcd/etcd-crc" Dec 03 00:06:57 crc kubenswrapper[4953]: I1203 00:06:57.301621 4953 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 00:06:57 crc kubenswrapper[4953]: I1203 00:06:57.302031 4953 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 00:06:57 crc kubenswrapper[4953]: I1203 00:06:57.304254 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:06:57 crc kubenswrapper[4953]: I1203 00:06:57.304334 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:06:57 crc kubenswrapper[4953]: I1203 00:06:57.304362 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:06:57 crc kubenswrapper[4953]: I1203 00:06:57.415174 4953 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 00:06:57 crc kubenswrapper[4953]: I1203 00:06:57.416893 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:06:57 crc kubenswrapper[4953]: I1203 00:06:57.417007 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:06:57 crc kubenswrapper[4953]: I1203 00:06:57.417028 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:06:57 crc kubenswrapper[4953]: I1203 00:06:57.947408 4953 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-etcd/etcd-crc" Dec 03 00:06:58 crc kubenswrapper[4953]: I1203 00:06:58.417914 4953 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 00:06:58 crc kubenswrapper[4953]: I1203 00:06:58.419009 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:06:58 crc kubenswrapper[4953]: I1203 00:06:58.419056 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:06:58 crc kubenswrapper[4953]: I1203 00:06:58.419073 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:06:59 crc kubenswrapper[4953]: I1203 00:06:59.421859 4953 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 00:06:59 crc kubenswrapper[4953]: I1203 00:06:59.423804 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:06:59 crc kubenswrapper[4953]: I1203 00:06:59.423833 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:06:59 crc kubenswrapper[4953]: I1203 00:06:59.423845 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:06:59 crc kubenswrapper[4953]: I1203 00:06:59.863849 4953 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 03 00:06:59 crc kubenswrapper[4953]: I1203 00:06:59.863963 4953 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 03 00:07:00 crc kubenswrapper[4953]: I1203 00:07:00.640583 4953 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 03 00:07:00 crc kubenswrapper[4953]: I1203 00:07:00.640820 4953 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 00:07:00 crc kubenswrapper[4953]: I1203 00:07:00.643271 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:00 crc kubenswrapper[4953]: I1203 00:07:00.643332 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:00 crc kubenswrapper[4953]: I1203 00:07:00.643350 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:01 crc kubenswrapper[4953]: E1203 00:07:01.408740 4953 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Dec 03 00:07:01 crc kubenswrapper[4953]: I1203 00:07:01.860711 4953 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 03 00:07:01 crc kubenswrapper[4953]: I1203 00:07:01.861574 4953 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 00:07:01 crc kubenswrapper[4953]: I1203 00:07:01.863240 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:01 crc kubenswrapper[4953]: I1203 00:07:01.863332 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:01 crc kubenswrapper[4953]: I1203 00:07:01.863370 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:01 crc kubenswrapper[4953]: I1203 00:07:01.868171 4953 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 03 00:07:02 crc kubenswrapper[4953]: I1203 00:07:02.430683 4953 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 00:07:02 crc kubenswrapper[4953]: I1203 00:07:02.432508 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:02 crc kubenswrapper[4953]: I1203 00:07:02.432597 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:02 crc kubenswrapper[4953]: I1203 00:07:02.432625 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:02 crc kubenswrapper[4953]: I1203 00:07:02.443614 4953 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 03 00:07:03 crc kubenswrapper[4953]: I1203 00:07:03.273825 4953 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": net/http: TLS handshake timeout Dec 03 00:07:03 crc kubenswrapper[4953]: I1203 00:07:03.433328 4953 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 00:07:03 crc kubenswrapper[4953]: I1203 00:07:03.434515 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:03 crc kubenswrapper[4953]: I1203 00:07:03.434585 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:03 crc kubenswrapper[4953]: I1203 00:07:03.434604 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:04 crc kubenswrapper[4953]: I1203 00:07:04.269857 4953 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Dec 03 00:07:04 crc kubenswrapper[4953]: I1203 00:07:04.269919 4953 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Dec 03 00:07:04 crc kubenswrapper[4953]: I1203 00:07:04.286469 4953 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Dec 03 00:07:04 crc kubenswrapper[4953]: I1203 00:07:04.286550 4953 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Dec 03 00:07:04 crc kubenswrapper[4953]: I1203 00:07:04.445836 4953 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[+]ping ok Dec 03 00:07:04 crc kubenswrapper[4953]: [+]log ok Dec 03 00:07:04 crc kubenswrapper[4953]: [+]etcd ok Dec 03 00:07:04 crc kubenswrapper[4953]: [+]poststarthook/openshift.io-openshift-apiserver-reachable ok Dec 03 00:07:04 crc kubenswrapper[4953]: [+]poststarthook/openshift.io-oauth-apiserver-reachable ok Dec 03 00:07:04 crc kubenswrapper[4953]: [+]poststarthook/start-apiserver-admission-initializer ok Dec 03 00:07:04 crc kubenswrapper[4953]: [+]poststarthook/quota.openshift.io-clusterquotamapping ok Dec 03 00:07:04 crc kubenswrapper[4953]: [+]poststarthook/openshift.io-api-request-count-filter ok Dec 03 00:07:04 crc kubenswrapper[4953]: [+]poststarthook/openshift.io-startkubeinformers ok Dec 03 00:07:04 crc kubenswrapper[4953]: [+]poststarthook/generic-apiserver-start-informers ok Dec 03 00:07:04 crc kubenswrapper[4953]: [+]poststarthook/priority-and-fairness-config-consumer ok Dec 03 00:07:04 crc kubenswrapper[4953]: [+]poststarthook/priority-and-fairness-filter ok Dec 03 00:07:04 crc kubenswrapper[4953]: [+]poststarthook/storage-object-count-tracker-hook ok Dec 03 00:07:04 crc kubenswrapper[4953]: [+]poststarthook/start-apiextensions-informers ok Dec 03 00:07:04 crc kubenswrapper[4953]: [-]poststarthook/start-apiextensions-controllers failed: reason withheld Dec 03 00:07:04 crc kubenswrapper[4953]: [-]poststarthook/crd-informer-synced failed: reason withheld Dec 03 00:07:04 crc kubenswrapper[4953]: [+]poststarthook/start-system-namespaces-controller ok Dec 03 00:07:04 crc kubenswrapper[4953]: [+]poststarthook/start-cluster-authentication-info-controller ok Dec 03 00:07:04 crc kubenswrapper[4953]: [+]poststarthook/start-kube-apiserver-identity-lease-controller ok Dec 03 00:07:04 crc kubenswrapper[4953]: [+]poststarthook/start-kube-apiserver-identity-lease-garbage-collector ok Dec 03 00:07:04 crc kubenswrapper[4953]: [+]poststarthook/start-legacy-token-tracking-controller ok Dec 03 00:07:04 crc kubenswrapper[4953]: [+]poststarthook/start-service-ip-repair-controllers ok Dec 03 00:07:04 crc kubenswrapper[4953]: [-]poststarthook/rbac/bootstrap-roles failed: reason withheld Dec 03 00:07:04 crc kubenswrapper[4953]: [-]poststarthook/scheduling/bootstrap-system-priority-classes failed: reason withheld Dec 03 00:07:04 crc kubenswrapper[4953]: [+]poststarthook/priority-and-fairness-config-producer ok Dec 03 00:07:04 crc kubenswrapper[4953]: [+]poststarthook/bootstrap-controller ok Dec 03 00:07:04 crc kubenswrapper[4953]: [+]poststarthook/aggregator-reload-proxy-client-cert ok Dec 03 00:07:04 crc kubenswrapper[4953]: [+]poststarthook/start-kube-aggregator-informers ok Dec 03 00:07:04 crc kubenswrapper[4953]: [+]poststarthook/apiservice-status-local-available-controller ok Dec 03 00:07:04 crc kubenswrapper[4953]: [+]poststarthook/apiservice-status-remote-available-controller ok Dec 03 00:07:04 crc kubenswrapper[4953]: [+]poststarthook/apiservice-registration-controller ok Dec 03 00:07:04 crc kubenswrapper[4953]: [+]poststarthook/apiservice-wait-for-first-sync ok Dec 03 00:07:04 crc kubenswrapper[4953]: [+]poststarthook/apiservice-discovery-controller ok Dec 03 00:07:04 crc kubenswrapper[4953]: [+]poststarthook/kube-apiserver-autoregistration ok Dec 03 00:07:04 crc kubenswrapper[4953]: [+]autoregister-completion ok Dec 03 00:07:04 crc kubenswrapper[4953]: [+]poststarthook/apiservice-openapi-controller ok Dec 03 00:07:04 crc kubenswrapper[4953]: [+]poststarthook/apiservice-openapiv3-controller ok Dec 03 00:07:04 crc kubenswrapper[4953]: livez check failed Dec 03 00:07:04 crc kubenswrapper[4953]: I1203 00:07:04.445914 4953 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 03 00:07:07 crc kubenswrapper[4953]: I1203 00:07:07.305785 4953 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-etcd/etcd-crc" Dec 03 00:07:07 crc kubenswrapper[4953]: I1203 00:07:07.306108 4953 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 00:07:07 crc kubenswrapper[4953]: I1203 00:07:07.307452 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:07 crc kubenswrapper[4953]: I1203 00:07:07.307546 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:07 crc kubenswrapper[4953]: I1203 00:07:07.307566 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:07 crc kubenswrapper[4953]: I1203 00:07:07.328760 4953 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-etcd/etcd-crc" Dec 03 00:07:07 crc kubenswrapper[4953]: I1203 00:07:07.444532 4953 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 00:07:07 crc kubenswrapper[4953]: I1203 00:07:07.445960 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:07 crc kubenswrapper[4953]: I1203 00:07:07.446033 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:07 crc kubenswrapper[4953]: I1203 00:07:07.446055 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:09 crc kubenswrapper[4953]: E1203 00:07:09.279603 4953 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": context deadline exceeded" interval="3.2s" Dec 03 00:07:09 crc kubenswrapper[4953]: I1203 00:07:09.282157 4953 trace.go:236] Trace[1632028310]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (03-Dec-2025 00:06:54.992) (total time: 14289ms): Dec 03 00:07:09 crc kubenswrapper[4953]: Trace[1632028310]: ---"Objects listed" error: 14289ms (00:07:09.281) Dec 03 00:07:09 crc kubenswrapper[4953]: Trace[1632028310]: [14.289622943s] [14.289622943s] END Dec 03 00:07:09 crc kubenswrapper[4953]: I1203 00:07:09.282195 4953 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Dec 03 00:07:09 crc kubenswrapper[4953]: I1203 00:07:09.283204 4953 trace.go:236] Trace[1166222559]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (03-Dec-2025 00:06:55.157) (total time: 14125ms): Dec 03 00:07:09 crc kubenswrapper[4953]: Trace[1166222559]: ---"Objects listed" error: 14125ms (00:07:09.283) Dec 03 00:07:09 crc kubenswrapper[4953]: Trace[1166222559]: [14.125661763s] [14.125661763s] END Dec 03 00:07:09 crc kubenswrapper[4953]: I1203 00:07:09.283247 4953 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Dec 03 00:07:09 crc kubenswrapper[4953]: I1203 00:07:09.285250 4953 reconstruct.go:205] "DevicePaths of reconstructed volumes updated" Dec 03 00:07:09 crc kubenswrapper[4953]: I1203 00:07:09.285502 4953 trace.go:236] Trace[1916839487]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (03-Dec-2025 00:06:55.018) (total time: 14267ms): Dec 03 00:07:09 crc kubenswrapper[4953]: Trace[1916839487]: ---"Objects listed" error: 14266ms (00:07:09.285) Dec 03 00:07:09 crc kubenswrapper[4953]: Trace[1916839487]: [14.267106053s] [14.267106053s] END Dec 03 00:07:09 crc kubenswrapper[4953]: I1203 00:07:09.285669 4953 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Dec 03 00:07:09 crc kubenswrapper[4953]: I1203 00:07:09.286671 4953 trace.go:236] Trace[1982308102]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (03-Dec-2025 00:06:54.367) (total time: 14919ms): Dec 03 00:07:09 crc kubenswrapper[4953]: Trace[1982308102]: ---"Objects listed" error: 14918ms (00:07:09.286) Dec 03 00:07:09 crc kubenswrapper[4953]: Trace[1982308102]: [14.919074137s] [14.919074137s] END Dec 03 00:07:09 crc kubenswrapper[4953]: I1203 00:07:09.287521 4953 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Dec 03 00:07:09 crc kubenswrapper[4953]: E1203 00:07:09.287862 4953 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes \"crc\" is forbidden: autoscaling.openshift.io/ManagedNode infra config cache not synchronized" node="crc" Dec 03 00:07:09 crc kubenswrapper[4953]: I1203 00:07:09.342144 4953 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:53954->192.168.126.11:17697: read: connection reset by peer" start-of-body= Dec 03 00:07:09 crc kubenswrapper[4953]: I1203 00:07:09.342225 4953 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:53954->192.168.126.11:17697: read: connection reset by peer" Dec 03 00:07:09 crc kubenswrapper[4953]: I1203 00:07:09.343117 4953 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Liveness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:53966->192.168.126.11:17697: read: connection reset by peer" start-of-body= Dec 03 00:07:09 crc kubenswrapper[4953]: I1203 00:07:09.343656 4953 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:53966->192.168.126.11:17697: read: connection reset by peer" Dec 03 00:07:09 crc kubenswrapper[4953]: I1203 00:07:09.384062 4953 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 03 00:07:09 crc kubenswrapper[4953]: I1203 00:07:09.390520 4953 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 03 00:07:09 crc kubenswrapper[4953]: I1203 00:07:09.441422 4953 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 00:07:09 crc kubenswrapper[4953]: I1203 00:07:09.442277 4953 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" start-of-body= Dec 03 00:07:09 crc kubenswrapper[4953]: I1203 00:07:09.442444 4953 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" Dec 03 00:07:09 crc kubenswrapper[4953]: I1203 00:07:09.446513 4953 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 00:07:09 crc kubenswrapper[4953]: I1203 00:07:09.450804 4953 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Dec 03 00:07:09 crc kubenswrapper[4953]: I1203 00:07:09.452932 4953 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="761973393a93c59c23e69f41b8ac0d16abc35abd61ecbc6d51a42d94cd664233" exitCode=255 Dec 03 00:07:09 crc kubenswrapper[4953]: I1203 00:07:09.453664 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"761973393a93c59c23e69f41b8ac0d16abc35abd61ecbc6d51a42d94cd664233"} Dec 03 00:07:09 crc kubenswrapper[4953]: E1203 00:07:09.459164 4953 kubelet.go:1929] "Failed creating a mirror pod for" err="pods \"kube-controller-manager-crc\" already exists" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.280829 4953 apiserver.go:52] "Watching apiserver" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.286282 4953 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.286782 4953 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-network-operator/iptables-alerter-4ln5h","openshift-network-operator/network-operator-58b4c7f79c-55gtf","openshift-kube-apiserver/kube-apiserver-crc","openshift-kube-controller-manager/kube-controller-manager-crc","openshift-network-console/networking-console-plugin-85b44fc459-gdk6g","openshift-network-diagnostics/network-check-source-55646444c4-trplf","openshift-network-diagnostics/network-check-target-xd92c","openshift-network-node-identity/network-node-identity-vrzqb"] Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.287359 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.287515 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 00:07:10 crc kubenswrapper[4953]: E1203 00:07:10.287666 4953 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.287684 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 00:07:10 crc kubenswrapper[4953]: E1203 00:07:10.287753 4953 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.288054 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.289350 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 00:07:10 crc kubenswrapper[4953]: E1203 00:07:10.289639 4953 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.289522 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.292407 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.293508 4953 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.293903 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.294629 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.298033 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.298261 4953 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.298469 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.298710 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.299825 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.333880 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:10Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.355080 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:10Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.370951 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f94f8c51-c665-4ca3-bf17-7a04ea5c7ade\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72c5cd7dc1c3a8adfedab6ba43a59d80b3efe65f2bb56be95e76cc86a409dcf3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2810ffc1b141fa7da811406053db9d72c708edf5f2bb1dd75e4d52ab35fd24c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c67b4277b20a9f1287f50ef82d0785c9d02543a8e7ed26e0413eae332b866db7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://258d78e1ad9d3dcda6976909ae3fe71aecead5b995fc66b3fd6e2722d80b5ab8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:06:51Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.375685 4953 desired_state_of_world_populator.go:154] "Finished populating initial desired state of world" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.391922 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.392009 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.392036 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.392063 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.392084 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.392107 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.392131 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.392151 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.392173 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.392195 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.392216 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.392240 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.392265 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.392299 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.392332 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.392361 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.392388 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.392422 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.392450 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.392473 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.392496 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.392519 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.392542 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.392570 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.392592 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.392613 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.392634 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.392637 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" (OuterVolumeSpecName: "kube-api-access-mg5zb") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "kube-api-access-mg5zb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.392656 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.392760 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.392758 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" (OuterVolumeSpecName: "kube-api-access-gf66m") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "kube-api-access-gf66m". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.392815 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.392842 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.392880 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.392898 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.392912 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.392942 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.392960 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.393021 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.393038 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.393054 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.393086 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.393105 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.393120 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.393152 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.393170 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.393190 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.393208 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.393242 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.393265 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.393283 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.393317 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.393334 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.393359 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.393418 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.393439 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.393458 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.393514 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.393531 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.393550 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.393581 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.393597 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.393612 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") pod \"44663579-783b-4372-86d6-acf235a62d72\" (UID: \"44663579-783b-4372-86d6-acf235a62d72\") " Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.393629 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.393662 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.393684 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.393701 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.393746 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.393765 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.393783 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.393821 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.393840 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.393863 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") pod \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\" (UID: \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\") " Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.393898 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.393919 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.394023 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.394044 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.394063 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.394099 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.394119 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.394139 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.394174 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.394192 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.394211 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.394247 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.394264 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.394283 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.394300 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.394336 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.394353 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.394369 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.394493 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.394513 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.394529 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.394548 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.394622 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.394640 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.394658 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.394693 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.394708 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.394726 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.394741 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.394773 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.394789 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.395033 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.395058 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.395074 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.395112 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.395130 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") pod \"49ef4625-1d3a-4a9f-b595-c2433d32326d\" (UID: \"49ef4625-1d3a-4a9f-b595-c2433d32326d\") " Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.395149 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.395166 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.395201 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.395218 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.395236 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.395252 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.395286 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.395303 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.395318 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.395367 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.395388 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.395407 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.395445 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.395465 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.395480 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.395496 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.395529 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") pod \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\" (UID: \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\") " Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.395545 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.395560 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.395594 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.395614 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.395633 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.395651 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.395689 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.395705 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.395721 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.395757 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.395774 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.395854 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.395880 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.395921 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.395956 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.396000 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.396018 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.396034 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.396069 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.396088 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.396106 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.396125 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.396159 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.396179 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.396198 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.396231 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.396248 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.396265 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.396353 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.396392 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.396410 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.396428 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.396446 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.396537 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.396558 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.396576 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.396772 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.396791 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.396812 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.396854 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.396872 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.396891 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.396921 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.396938 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.396957 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.396999 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.397017 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.397035 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.397052 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.397085 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.397105 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.397149 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.397171 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.397189 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.397207 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.397240 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.397257 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.397277 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.397308 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.397324 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.397342 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.397359 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.397394 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.397412 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.397431 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.397465 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.397494 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.397541 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.397563 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.397581 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.397615 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.397659 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.397699 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.397724 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.397748 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.397794 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.397819 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.397862 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.397881 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.397903 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.392963 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.393052 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" (OuterVolumeSpecName: "config") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.397940 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.397957 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.393052 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" (OuterVolumeSpecName: "config") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.393070 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" (OuterVolumeSpecName: "config") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.398026 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.393111 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.393127 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.393212 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.393312 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" (OuterVolumeSpecName: "machine-approver-tls") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "machine-approver-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.393464 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.393606 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.393601 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.393631 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.393694 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" (OuterVolumeSpecName: "kube-api-access-d6qdx") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "kube-api-access-d6qdx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.393836 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.393902 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" (OuterVolumeSpecName: "kube-api-access-fcqwp") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "kube-api-access-fcqwp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.393938 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.394016 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.394055 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" (OuterVolumeSpecName: "kube-api-access-dbsvg") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "kube-api-access-dbsvg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.394171 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" (OuterVolumeSpecName: "config") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.394196 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" (OuterVolumeSpecName: "utilities") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.394203 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" (OuterVolumeSpecName: "kube-api-access-htfz6") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "kube-api-access-htfz6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.394283 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" (OuterVolumeSpecName: "certs") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.394288 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" (OuterVolumeSpecName: "webhook-certs") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "webhook-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.394473 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" (OuterVolumeSpecName: "config") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.394420 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" (OuterVolumeSpecName: "kube-api-access-bf2bz") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "kube-api-access-bf2bz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.394520 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" (OuterVolumeSpecName: "kube-api-access-s4n52") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "kube-api-access-s4n52". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.394555 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" (OuterVolumeSpecName: "kube-api-access-rnphk") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "kube-api-access-rnphk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.394599 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" (OuterVolumeSpecName: "config-volume") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.394820 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" (OuterVolumeSpecName: "kube-api-access-2d4wz") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "kube-api-access-2d4wz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.394855 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" (OuterVolumeSpecName: "config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.394934 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.395247 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" (OuterVolumeSpecName: "kube-api-access-nzwt7") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "kube-api-access-nzwt7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.395581 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" (OuterVolumeSpecName: "kube-api-access-x2m85") pod "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" (UID: "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d"). InnerVolumeSpecName "kube-api-access-x2m85". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.395986 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.396162 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.396226 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.396250 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.396652 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" (OuterVolumeSpecName: "kube-api-access-sb6h7") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "kube-api-access-sb6h7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.396795 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.396875 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" (OuterVolumeSpecName: "kube-api-access-cfbct") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "kube-api-access-cfbct". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.397110 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" (OuterVolumeSpecName: "kube-api-access-zgdk5") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "kube-api-access-zgdk5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.397181 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" (OuterVolumeSpecName: "kube-api-access-4d4hj") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "kube-api-access-4d4hj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.397361 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" (OuterVolumeSpecName: "image-registry-operator-tls") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "image-registry-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.397525 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.397744 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" (OuterVolumeSpecName: "kube-api-access-fqsjt") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "kube-api-access-fqsjt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.397901 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" (OuterVolumeSpecName: "kube-api-access-w9rds") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "kube-api-access-w9rds". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.398635 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" (OuterVolumeSpecName: "config") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.398736 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" (OuterVolumeSpecName: "ovn-control-plane-metrics-cert") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovn-control-plane-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.398071 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.398809 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.398925 4953 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") on node \"crc\" DevicePath \"\"" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.398943 4953 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.398965 4953 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") on node \"crc\" DevicePath \"\"" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.399000 4953 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.399012 4953 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") on node \"crc\" DevicePath \"\"" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.399023 4953 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") on node \"crc\" DevicePath \"\"" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.399034 4953 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") on node \"crc\" DevicePath \"\"" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.399046 4953 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.399057 4953 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.399067 4953 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") on node \"crc\" DevicePath \"\"" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.399078 4953 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") on node \"crc\" DevicePath \"\"" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.399089 4953 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") on node \"crc\" DevicePath \"\"" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.399101 4953 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") on node \"crc\" DevicePath \"\"" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.399099 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" (OuterVolumeSpecName: "multus-daemon-config") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "multus-daemon-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.399115 4953 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") on node \"crc\" DevicePath \"\"" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.399286 4953 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") on node \"crc\" DevicePath \"\"" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.399313 4953 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.399341 4953 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") on node \"crc\" DevicePath \"\"" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.399302 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" (OuterVolumeSpecName: "kube-api-access-pcxfs") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "kube-api-access-pcxfs". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.399397 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" (OuterVolumeSpecName: "available-featuregates") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "available-featuregates". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.399379 4953 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.399489 4953 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") on node \"crc\" DevicePath \"\"" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.399523 4953 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") on node \"crc\" DevicePath \"\"" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.399627 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" (OuterVolumeSpecName: "samples-operator-tls") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "samples-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.399705 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.399727 4953 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") on node \"crc\" DevicePath \"\"" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.399737 4953 swap_util.go:74] "error creating dir to test if tmpfs noswap is enabled. Assuming not supported" mount path="" error="stat /var/lib/kubelet/plugins/kubernetes.io/empty-dir: no such file or directory" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.399776 4953 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") on node \"crc\" DevicePath \"\"" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.399811 4953 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.399842 4953 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") on node \"crc\" DevicePath \"\"" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.399870 4953 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.399896 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" (OuterVolumeSpecName: "kube-api-access-vt5rc") pod "44663579-783b-4372-86d6-acf235a62d72" (UID: "44663579-783b-4372-86d6-acf235a62d72"). InnerVolumeSpecName "kube-api-access-vt5rc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.399903 4953 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") on node \"crc\" DevicePath \"\"" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.399993 4953 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.400010 4953 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") on node \"crc\" DevicePath \"\"" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.400012 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" (OuterVolumeSpecName: "kube-api-access-kfwg7") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "kube-api-access-kfwg7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.400028 4953 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") on node \"crc\" DevicePath \"\"" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.400047 4953 reconciler_common.go:293] "Volume detached for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") on node \"crc\" DevicePath \"\"" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.400061 4953 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") on node \"crc\" DevicePath \"\"" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.400075 4953 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") on node \"crc\" DevicePath \"\"" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.400089 4953 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.400102 4953 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.400116 4953 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") on node \"crc\" DevicePath \"\"" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.400130 4953 reconciler_common.go:293] "Volume detached for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") on node \"crc\" DevicePath \"\"" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.400141 4953 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") on node \"crc\" DevicePath \"\"" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.400152 4953 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") on node \"crc\" DevicePath \"\"" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.400165 4953 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") on node \"crc\" DevicePath \"\"" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.400177 4953 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") on node \"crc\" DevicePath \"\"" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.400190 4953 reconciler_common.go:293] "Volume detached for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") on node \"crc\" DevicePath \"\"" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.400203 4953 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") on node \"crc\" DevicePath \"\"" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.400216 4953 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") on node \"crc\" DevicePath \"\"" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.400228 4953 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") on node \"crc\" DevicePath \"\"" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.400241 4953 reconciler_common.go:293] "Volume detached for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") on node \"crc\" DevicePath \"\"" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.400256 4953 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.400268 4953 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") on node \"crc\" DevicePath \"\"" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.400286 4953 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") on node \"crc\" DevicePath \"\"" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.400300 4953 reconciler_common.go:293] "Volume detached for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") on node \"crc\" DevicePath \"\"" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.400383 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.400775 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" (OuterVolumeSpecName: "tmpfs") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "tmpfs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.401733 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" (OuterVolumeSpecName: "config") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.402078 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" (OuterVolumeSpecName: "console-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.402730 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.402965 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" (OuterVolumeSpecName: "cni-sysctl-allowlist") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-sysctl-allowlist". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.403051 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" (OuterVolumeSpecName: "kube-api-access-w7l8j") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "kube-api-access-w7l8j". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.403297 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.403924 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.403933 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.404429 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.404616 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.404623 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" (OuterVolumeSpecName: "kube-api-access-d4lsv") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "kube-api-access-d4lsv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.404801 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.405009 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" (OuterVolumeSpecName: "kube-api-access-9xfj7") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "kube-api-access-9xfj7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.405261 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.405656 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.405745 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" (OuterVolumeSpecName: "package-server-manager-serving-cert") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "package-server-manager-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.406041 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.406277 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.407040 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.407233 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.407652 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" (OuterVolumeSpecName: "apiservice-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "apiservice-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.407684 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" (OuterVolumeSpecName: "config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.407829 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.408088 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" (OuterVolumeSpecName: "kube-api-access-tk88c") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "kube-api-access-tk88c". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.408171 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.408294 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.408417 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" (OuterVolumeSpecName: "utilities") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.408505 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" (OuterVolumeSpecName: "kube-api-access-zkvpv") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "kube-api-access-zkvpv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.409122 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" (OuterVolumeSpecName: "config") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.409835 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" (OuterVolumeSpecName: "control-plane-machine-set-operator-tls") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "control-plane-machine-set-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.410091 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" (OuterVolumeSpecName: "kube-api-access-6g6sz") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "kube-api-access-6g6sz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.410349 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.410671 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.410836 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.410901 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" (OuterVolumeSpecName: "serviceca") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "serviceca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.411415 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.411851 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" (OuterVolumeSpecName: "config") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.412162 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.412476 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.413664 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.413794 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" (OuterVolumeSpecName: "node-bootstrap-token") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "node-bootstrap-token". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.413802 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" (OuterVolumeSpecName: "images") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.414020 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" (OuterVolumeSpecName: "kube-api-access-ngvvp") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "kube-api-access-ngvvp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.414308 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" (OuterVolumeSpecName: "kube-api-access-2w9zh") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "kube-api-access-2w9zh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.414381 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" (OuterVolumeSpecName: "utilities") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.414409 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" (OuterVolumeSpecName: "image-import-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "image-import-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.414517 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.414579 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" (OuterVolumeSpecName: "kube-api-access-lz9wn") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "kube-api-access-lz9wn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.414890 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" (OuterVolumeSpecName: "kube-api-access-pjr6v") pod "49ef4625-1d3a-4a9f-b595-c2433d32326d" (UID: "49ef4625-1d3a-4a9f-b595-c2433d32326d"). InnerVolumeSpecName "kube-api-access-pjr6v". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.415067 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" (OuterVolumeSpecName: "kube-api-access-qg5z5") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "kube-api-access-qg5z5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.415429 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" (OuterVolumeSpecName: "stats-auth") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "stats-auth". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.415457 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.415465 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" (OuterVolumeSpecName: "kube-api-access-wxkg8") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "kube-api-access-wxkg8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.415516 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.415673 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.416308 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" (OuterVolumeSpecName: "mcc-auth-proxy-config") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "mcc-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.416525 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" (OuterVolumeSpecName: "images") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.416544 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.416896 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" (OuterVolumeSpecName: "utilities") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.416927 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" (OuterVolumeSpecName: "signing-cabundle") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-cabundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.417141 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" (OuterVolumeSpecName: "kube-api-access-xcphl") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "kube-api-access-xcphl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.417266 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 03 00:07:10 crc kubenswrapper[4953]: E1203 00:07:10.417412 4953 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.417438 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" (OuterVolumeSpecName: "audit") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "audit". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.417457 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.417495 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" (OuterVolumeSpecName: "kube-api-access-249nr") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "kube-api-access-249nr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:07:10 crc kubenswrapper[4953]: E1203 00:07:10.417519 4953 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.417539 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" (OuterVolumeSpecName: "config") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.417607 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.418107 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" (OuterVolumeSpecName: "kube-api-access-pj782") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "kube-api-access-pj782". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.417411 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.418385 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" (OuterVolumeSpecName: "kube-api-access-x7zkh") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "kube-api-access-x7zkh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.419550 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.419570 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" (OuterVolumeSpecName: "cert") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.419937 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" (OuterVolumeSpecName: "service-ca") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.423645 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" (OuterVolumeSpecName: "kube-api-access-jkwtn") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "kube-api-access-jkwtn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:07:10 crc kubenswrapper[4953]: E1203 00:07:10.417482 4953 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-03 00:07:10.917463749 +0000 UTC m=+19.901744532 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 03 00:07:10 crc kubenswrapper[4953]: E1203 00:07:10.428991 4953 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-03 00:07:10.928942873 +0000 UTC m=+19.913223656 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.429580 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" (OuterVolumeSpecName: "config") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.429863 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.430519 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" (OuterVolumeSpecName: "kube-api-access-279lb") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "kube-api-access-279lb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.430767 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" (OuterVolumeSpecName: "mcd-auth-proxy-config") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "mcd-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.430956 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" (OuterVolumeSpecName: "webhook-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "webhook-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.431010 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" (OuterVolumeSpecName: "kube-api-access-v47cf") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "kube-api-access-v47cf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.431204 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" (OuterVolumeSpecName: "kube-api-access-w4xd4") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "kube-api-access-w4xd4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.431331 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" (OuterVolumeSpecName: "config") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.431412 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.431348 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" (OuterVolumeSpecName: "kube-api-access-6ccd8") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "kube-api-access-6ccd8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.431552 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" (OuterVolumeSpecName: "kube-api-access-qs4fp") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "kube-api-access-qs4fp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.431819 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.431888 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.432018 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" (OuterVolumeSpecName: "kube-api-access-lzf88") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "kube-api-access-lzf88". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.432086 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.432106 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.432308 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.432328 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.432397 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" (OuterVolumeSpecName: "kube-api-access-x4zgh") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "kube-api-access-x4zgh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.432076 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.433120 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.435333 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.435649 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.437528 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.438057 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" (OuterVolumeSpecName: "default-certificate") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "default-certificate". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.438425 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" (OuterVolumeSpecName: "kube-api-access-xcgwh") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "kube-api-access-xcgwh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.438521 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.438670 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.438967 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:07:10 crc kubenswrapper[4953]: E1203 00:07:10.439543 4953 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 00:07:10.939512035 +0000 UTC m=+19.923792828 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.439681 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.440203 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.441919 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" (OuterVolumeSpecName: "client-ca") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.442250 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.443432 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.443544 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.443850 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.445178 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" (OuterVolumeSpecName: "machine-api-operator-tls") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "machine-api-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.445305 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.445458 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" (OuterVolumeSpecName: "config") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.445474 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.446243 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.435581 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:10Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 00:07:10 crc kubenswrapper[4953]: E1203 00:07:10.449226 4953 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 03 00:07:10 crc kubenswrapper[4953]: E1203 00:07:10.449247 4953 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 03 00:07:10 crc kubenswrapper[4953]: E1203 00:07:10.449257 4953 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 00:07:10 crc kubenswrapper[4953]: E1203 00:07:10.449301 4953 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-03 00:07:10.949290887 +0000 UTC m=+19.933571670 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.450344 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.451107 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.451208 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.452268 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.452798 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.454861 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.457255 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:07:10 crc kubenswrapper[4953]: E1203 00:07:10.460268 4953 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 03 00:07:10 crc kubenswrapper[4953]: E1203 00:07:10.460295 4953 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 03 00:07:10 crc kubenswrapper[4953]: E1203 00:07:10.460311 4953 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 00:07:10 crc kubenswrapper[4953]: E1203 00:07:10.460368 4953 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-03 00:07:10.96035006 +0000 UTC m=+19.944630843 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.460688 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.460761 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" (OuterVolumeSpecName: "service-ca") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.460794 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" (OuterVolumeSpecName: "etcd-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.460895 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" (OuterVolumeSpecName: "kube-api-access-8tdtz") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "kube-api-access-8tdtz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.461168 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" (OuterVolumeSpecName: "kube-api-access-jhbk2") pod "bd23aa5c-e532-4e53-bccf-e79f130c5ae8" (UID: "bd23aa5c-e532-4e53-bccf-e79f130c5ae8"). InnerVolumeSpecName "kube-api-access-jhbk2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.461417 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" (OuterVolumeSpecName: "etcd-service-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.462828 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.463961 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.464530 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" (OuterVolumeSpecName: "signing-key") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:07:10 crc kubenswrapper[4953]: E1203 00:07:10.464855 4953 kubelet.go:1929] "Failed creating a mirror pod for" err="pods \"kube-apiserver-crc\" already exists" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.464912 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.464925 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.465117 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.465322 4953 scope.go:117] "RemoveContainer" containerID="761973393a93c59c23e69f41b8ac0d16abc35abd61ecbc6d51a42d94cd664233" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.465653 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" (OuterVolumeSpecName: "client-ca") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.465916 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cf821840-5353-4620-8f73-99b56b371ab0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cae5e2994fcc24dc9b2ea73e79572615b3311edc42642b7f278fb84123989b23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://29e4b9d1b70a93fb133e0759ad148ae133766a66efeae3ff06a21dfae0e805ff\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dec43536772aa7c3f5a662fa7f3bd73b1cae2791ad454f85792287828d0ffeed\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://761973393a93c59c23e69f41b8ac0d16abc35abd61ecbc6d51a42d94cd664233\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c143d4232fb8bf7ef5999a664e431a00ab0db030f4d31a28d0bb21e00e38581\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:53Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72af6c27296eae2e275d455795c68f0d272869d63642b8353b17d466d0c901cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://72af6c27296eae2e275d455795c68f0d272869d63642b8353b17d466d0c901cf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:06:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:06:51Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.467451 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.467993 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" (OuterVolumeSpecName: "kube-api-access-7c4vf") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "kube-api-access-7c4vf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.468406 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" (OuterVolumeSpecName: "kube-api-access-mnrrd") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "kube-api-access-mnrrd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.469207 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.469703 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.479463 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.481241 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:10Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.483765 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.493612 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.493795 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.501668 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.501708 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.501760 4953 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") on node \"crc\" DevicePath \"\"" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.501772 4953 reconciler_common.go:293] "Volume detached for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.501785 4953 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.501794 4953 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.501805 4953 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") on node \"crc\" DevicePath \"\"" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.501814 4953 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") on node \"crc\" DevicePath \"\"" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.501827 4953 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.501839 4953 reconciler_common.go:293] "Volume detached for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.501851 4953 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") on node \"crc\" DevicePath \"\"" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.501860 4953 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.501869 4953 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") on node \"crc\" DevicePath \"\"" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.501878 4953 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") on node \"crc\" DevicePath \"\"" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.501888 4953 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") on node \"crc\" DevicePath \"\"" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.501898 4953 reconciler_common.go:293] "Volume detached for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") on node \"crc\" DevicePath \"\"" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.501907 4953 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") on node \"crc\" DevicePath \"\"" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.501917 4953 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") on node \"crc\" DevicePath \"\"" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.501926 4953 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.501934 4953 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.501943 4953 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.501954 4953 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.501963 4953 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") on node \"crc\" DevicePath \"\"" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.501984 4953 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.501993 4953 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.502002 4953 reconciler_common.go:293] "Volume detached for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") on node \"crc\" DevicePath \"\"" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.502010 4953 reconciler_common.go:293] "Volume detached for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") on node \"crc\" DevicePath \"\"" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.502020 4953 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.502037 4953 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") on node \"crc\" DevicePath \"\"" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.502052 4953 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") on node \"crc\" DevicePath \"\"" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.502063 4953 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.502076 4953 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") on node \"crc\" DevicePath \"\"" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.502087 4953 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") on node \"crc\" DevicePath \"\"" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.502097 4953 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.502108 4953 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.502118 4953 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") on node \"crc\" DevicePath \"\"" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.502131 4953 reconciler_common.go:293] "Volume detached for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") on node \"crc\" DevicePath \"\"" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.502143 4953 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") on node \"crc\" DevicePath \"\"" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.502154 4953 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.502166 4953 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.502178 4953 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.502190 4953 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") on node \"crc\" DevicePath \"\"" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.502201 4953 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") on node \"crc\" DevicePath \"\"" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.502213 4953 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") on node \"crc\" DevicePath \"\"" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.502223 4953 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.502232 4953 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") on node \"crc\" DevicePath \"\"" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.502240 4953 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.502248 4953 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") on node \"crc\" DevicePath \"\"" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.502256 4953 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.502266 4953 reconciler_common.go:293] "Volume detached for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") on node \"crc\" DevicePath \"\"" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.502275 4953 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.502283 4953 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") on node \"crc\" DevicePath \"\"" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.502293 4953 reconciler_common.go:293] "Volume detached for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") on node \"crc\" DevicePath \"\"" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.502303 4953 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.502312 4953 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") on node \"crc\" DevicePath \"\"" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.502323 4953 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.502332 4953 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") on node \"crc\" DevicePath \"\"" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.502342 4953 reconciler_common.go:293] "Volume detached for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") on node \"crc\" DevicePath \"\"" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.502352 4953 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") on node \"crc\" DevicePath \"\"" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.502362 4953 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.502371 4953 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") on node \"crc\" DevicePath \"\"" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.502381 4953 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") on node \"crc\" DevicePath \"\"" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.502390 4953 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") on node \"crc\" DevicePath \"\"" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.502399 4953 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") on node \"crc\" DevicePath \"\"" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.502411 4953 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.502421 4953 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.502431 4953 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") on node \"crc\" DevicePath \"\"" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.502440 4953 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") on node \"crc\" DevicePath \"\"" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.502449 4953 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") on node \"crc\" DevicePath \"\"" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.502458 4953 reconciler_common.go:293] "Volume detached for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") on node \"crc\" DevicePath \"\"" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.502466 4953 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") on node \"crc\" DevicePath \"\"" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.502475 4953 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.502483 4953 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") on node \"crc\" DevicePath \"\"" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.502492 4953 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") on node \"crc\" DevicePath \"\"" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.502500 4953 reconciler_common.go:293] "Volume detached for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") on node \"crc\" DevicePath \"\"" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.502508 4953 reconciler_common.go:293] "Volume detached for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") on node \"crc\" DevicePath \"\"" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.502516 4953 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.502524 4953 reconciler_common.go:293] "Volume detached for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") on node \"crc\" DevicePath \"\"" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.502532 4953 reconciler_common.go:293] "Volume detached for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") on node \"crc\" DevicePath \"\"" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.502542 4953 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.502550 4953 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") on node \"crc\" DevicePath \"\"" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.502559 4953 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") on node \"crc\" DevicePath \"\"" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.502566 4953 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.502575 4953 reconciler_common.go:293] "Volume detached for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") on node \"crc\" DevicePath \"\"" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.502584 4953 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") on node \"crc\" DevicePath \"\"" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.502594 4953 reconciler_common.go:293] "Volume detached for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.502603 4953 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.502613 4953 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.502622 4953 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.502634 4953 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.502644 4953 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") on node \"crc\" DevicePath \"\"" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.502653 4953 reconciler_common.go:293] "Volume detached for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") on node \"crc\" DevicePath \"\"" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.502661 4953 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.502671 4953 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.502680 4953 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") on node \"crc\" DevicePath \"\"" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.502689 4953 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") on node \"crc\" DevicePath \"\"" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.502698 4953 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") on node \"crc\" DevicePath \"\"" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.502706 4953 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.502715 4953 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.502723 4953 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") on node \"crc\" DevicePath \"\"" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.502732 4953 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") on node \"crc\" DevicePath \"\"" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.502752 4953 reconciler_common.go:293] "Volume detached for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") on node \"crc\" DevicePath \"\"" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.502761 4953 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") on node \"crc\" DevicePath \"\"" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.502769 4953 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.502778 4953 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") on node \"crc\" DevicePath \"\"" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.502786 4953 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.502795 4953 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.502803 4953 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") on node \"crc\" DevicePath \"\"" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.502812 4953 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") on node \"crc\" DevicePath \"\"" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.502821 4953 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.502831 4953 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") on node \"crc\" DevicePath \"\"" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.502840 4953 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") on node \"crc\" DevicePath \"\"" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.502849 4953 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") on node \"crc\" DevicePath \"\"" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.502857 4953 reconciler_common.go:293] "Volume detached for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") on node \"crc\" DevicePath \"\"" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.502866 4953 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.502874 4953 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.502883 4953 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") on node \"crc\" DevicePath \"\"" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.502891 4953 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") on node \"crc\" DevicePath \"\"" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.502900 4953 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") on node \"crc\" DevicePath \"\"" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.502908 4953 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") on node \"crc\" DevicePath \"\"" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.502916 4953 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") on node \"crc\" DevicePath \"\"" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.502925 4953 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.502934 4953 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") on node \"crc\" DevicePath \"\"" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.502946 4953 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.502954 4953 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") on node \"crc\" DevicePath \"\"" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.502963 4953 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") on node \"crc\" DevicePath \"\"" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.502987 4953 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") on node \"crc\" DevicePath \"\"" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.502996 4953 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") on node \"crc\" DevicePath \"\"" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.503005 4953 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.503013 4953 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") on node \"crc\" DevicePath \"\"" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.503028 4953 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") on node \"crc\" DevicePath \"\"" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.503037 4953 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") on node \"crc\" DevicePath \"\"" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.503045 4953 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.503053 4953 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") on node \"crc\" DevicePath \"\"" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.503062 4953 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.503071 4953 reconciler_common.go:293] "Volume detached for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") on node \"crc\" DevicePath \"\"" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.503081 4953 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.503090 4953 reconciler_common.go:293] "Volume detached for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") on node \"crc\" DevicePath \"\"" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.503099 4953 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") on node \"crc\" DevicePath \"\"" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.503107 4953 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.503116 4953 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.503125 4953 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") on node \"crc\" DevicePath \"\"" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.503133 4953 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") on node \"crc\" DevicePath \"\"" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.503143 4953 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.503151 4953 reconciler_common.go:293] "Volume detached for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") on node \"crc\" DevicePath \"\"" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.503162 4953 reconciler_common.go:293] "Volume detached for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") on node \"crc\" DevicePath \"\"" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.503304 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.503342 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.508586 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.512199 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:10Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.524212 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:10Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.534508 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:10Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.547041 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f94f8c51-c665-4ca3-bf17-7a04ea5c7ade\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72c5cd7dc1c3a8adfedab6ba43a59d80b3efe65f2bb56be95e76cc86a409dcf3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2810ffc1b141fa7da811406053db9d72c708edf5f2bb1dd75e4d52ab35fd24c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c67b4277b20a9f1287f50ef82d0785c9d02543a8e7ed26e0413eae332b866db7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://258d78e1ad9d3dcda6976909ae3fe71aecead5b995fc66b3fd6e2722d80b5ab8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:06:51Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.558475 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.571703 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cf821840-5353-4620-8f73-99b56b371ab0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cae5e2994fcc24dc9b2ea73e79572615b3311edc42642b7f278fb84123989b23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://29e4b9d1b70a93fb133e0759ad148ae133766a66efeae3ff06a21dfae0e805ff\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dec43536772aa7c3f5a662fa7f3bd73b1cae2791ad454f85792287828d0ffeed\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://761973393a93c59c23e69f41b8ac0d16abc35abd61ecbc6d51a42d94cd664233\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://761973393a93c59c23e69f41b8ac0d16abc35abd61ecbc6d51a42d94cd664233\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T00:07:09Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 00:07:03.713374 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 00:07:03.715052 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1528563048/tls.crt::/tmp/serving-cert-1528563048/tls.key\\\\\\\"\\\\nI1203 00:07:09.300052 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 00:07:09.304192 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 00:07:09.304255 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 00:07:09.304309 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 00:07:09.304323 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 00:07:09.313563 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 00:07:09.313597 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 00:07:09.313603 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 00:07:09.313609 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 00:07:09.313613 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 00:07:09.313618 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 00:07:09.313622 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 00:07:09.313709 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 00:07:09.327347 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T00:06:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c143d4232fb8bf7ef5999a664e431a00ab0db030f4d31a28d0bb21e00e38581\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:53Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72af6c27296eae2e275d455795c68f0d272869d63642b8353b17d466d0c901cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://72af6c27296eae2e275d455795c68f0d272869d63642b8353b17d466d0c901cf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:06:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:06:51Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.582597 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:10Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.593615 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.603923 4953 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.604603 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:10Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.610681 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.624814 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 03 00:07:10 crc kubenswrapper[4953]: W1203 00:07:10.626037 4953 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod37a5e44f_9a88_4405_be8a_b645485e7312.slice/crio-668498b97c57a5c7863587b20bb86e4251785fc8a31bc39658045a86caf4f6e9 WatchSource:0}: Error finding container 668498b97c57a5c7863587b20bb86e4251785fc8a31bc39658045a86caf4f6e9: Status 404 returned error can't find the container with id 668498b97c57a5c7863587b20bb86e4251785fc8a31bc39658045a86caf4f6e9 Dec 03 00:07:10 crc kubenswrapper[4953]: I1203 00:07:10.633445 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 03 00:07:10 crc kubenswrapper[4953]: W1203 00:07:10.638769 4953 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podef543e1b_8068_4ea3_b32a_61027b32e95d.slice/crio-49772fe1f3c67c7ceb49d288df16e3d1d444c51608d6562659d4e4637116c35c WatchSource:0}: Error finding container 49772fe1f3c67c7ceb49d288df16e3d1d444c51608d6562659d4e4637116c35c: Status 404 returned error can't find the container with id 49772fe1f3c67c7ceb49d288df16e3d1d444c51608d6562659d4e4637116c35c Dec 03 00:07:11 crc kubenswrapper[4953]: I1203 00:07:11.006294 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 00:07:11 crc kubenswrapper[4953]: I1203 00:07:11.006373 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 00:07:11 crc kubenswrapper[4953]: I1203 00:07:11.006400 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 00:07:11 crc kubenswrapper[4953]: I1203 00:07:11.006422 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 00:07:11 crc kubenswrapper[4953]: I1203 00:07:11.006444 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 00:07:11 crc kubenswrapper[4953]: E1203 00:07:11.006522 4953 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 03 00:07:11 crc kubenswrapper[4953]: E1203 00:07:11.006602 4953 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 00:07:12.006558141 +0000 UTC m=+20.990838924 (durationBeforeRetry 1s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 00:07:11 crc kubenswrapper[4953]: E1203 00:07:11.006655 4953 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-03 00:07:12.006644163 +0000 UTC m=+20.990924946 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 03 00:07:11 crc kubenswrapper[4953]: E1203 00:07:11.006833 4953 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 03 00:07:11 crc kubenswrapper[4953]: E1203 00:07:11.006859 4953 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 03 00:07:11 crc kubenswrapper[4953]: E1203 00:07:11.006876 4953 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 00:07:11 crc kubenswrapper[4953]: E1203 00:07:11.006922 4953 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-03 00:07:12.00691444 +0000 UTC m=+20.991195213 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 00:07:11 crc kubenswrapper[4953]: E1203 00:07:11.007015 4953 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 03 00:07:11 crc kubenswrapper[4953]: E1203 00:07:11.007053 4953 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-03 00:07:12.007038733 +0000 UTC m=+20.991319516 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 03 00:07:11 crc kubenswrapper[4953]: E1203 00:07:11.007017 4953 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 03 00:07:11 crc kubenswrapper[4953]: E1203 00:07:11.007085 4953 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 03 00:07:11 crc kubenswrapper[4953]: E1203 00:07:11.007096 4953 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 00:07:11 crc kubenswrapper[4953]: E1203 00:07:11.007149 4953 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-03 00:07:12.007142535 +0000 UTC m=+20.991423318 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 00:07:11 crc kubenswrapper[4953]: I1203 00:07:11.350958 4953 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="01ab3dd5-8196-46d0-ad33-122e2ca51def" path="/var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes" Dec 03 00:07:11 crc kubenswrapper[4953]: I1203 00:07:11.352037 4953 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" path="/var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes" Dec 03 00:07:11 crc kubenswrapper[4953]: I1203 00:07:11.353210 4953 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09efc573-dbb6-4249-bd59-9b87aba8dd28" path="/var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes" Dec 03 00:07:11 crc kubenswrapper[4953]: I1203 00:07:11.353834 4953 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b574797-001e-440a-8f4e-c0be86edad0f" path="/var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes" Dec 03 00:07:11 crc kubenswrapper[4953]: I1203 00:07:11.354784 4953 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b78653f-4ff9-4508-8672-245ed9b561e3" path="/var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes" Dec 03 00:07:11 crc kubenswrapper[4953]: I1203 00:07:11.355310 4953 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1386a44e-36a2-460c-96d0-0359d2b6f0f5" path="/var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes" Dec 03 00:07:11 crc kubenswrapper[4953]: I1203 00:07:11.355914 4953 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1bf7eb37-55a3-4c65-b768-a94c82151e69" path="/var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes" Dec 03 00:07:11 crc kubenswrapper[4953]: I1203 00:07:11.356818 4953 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d611f23-29be-4491-8495-bee1670e935f" path="/var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes" Dec 03 00:07:11 crc kubenswrapper[4953]: I1203 00:07:11.357447 4953 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20b0d48f-5fd6-431c-a545-e3c800c7b866" path="/var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/volumes" Dec 03 00:07:11 crc kubenswrapper[4953]: I1203 00:07:11.358349 4953 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" path="/var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes" Dec 03 00:07:11 crc kubenswrapper[4953]: I1203 00:07:11.358888 4953 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="22c825df-677d-4ca6-82db-3454ed06e783" path="/var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes" Dec 03 00:07:11 crc kubenswrapper[4953]: I1203 00:07:11.359912 4953 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25e176fe-21b4-4974-b1ed-c8b94f112a7f" path="/var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes" Dec 03 00:07:11 crc kubenswrapper[4953]: I1203 00:07:11.360424 4953 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" path="/var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes" Dec 03 00:07:11 crc kubenswrapper[4953]: I1203 00:07:11.360945 4953 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31d8b7a1-420e-4252-a5b7-eebe8a111292" path="/var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes" Dec 03 00:07:11 crc kubenswrapper[4953]: I1203 00:07:11.361965 4953 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ab1a177-2de0-46d9-b765-d0d0649bb42e" path="/var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/volumes" Dec 03 00:07:11 crc kubenswrapper[4953]: I1203 00:07:11.362525 4953 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" path="/var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes" Dec 03 00:07:11 crc kubenswrapper[4953]: I1203 00:07:11.367483 4953 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="43509403-f426-496e-be36-56cef71462f5" path="/var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes" Dec 03 00:07:11 crc kubenswrapper[4953]: I1203 00:07:11.368189 4953 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44663579-783b-4372-86d6-acf235a62d72" path="/var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/volumes" Dec 03 00:07:11 crc kubenswrapper[4953]: I1203 00:07:11.370707 4953 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="496e6271-fb68-4057-954e-a0d97a4afa3f" path="/var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes" Dec 03 00:07:11 crc kubenswrapper[4953]: I1203 00:07:11.371969 4953 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" path="/var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes" Dec 03 00:07:11 crc kubenswrapper[4953]: I1203 00:07:11.373875 4953 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49ef4625-1d3a-4a9f-b595-c2433d32326d" path="/var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/volumes" Dec 03 00:07:11 crc kubenswrapper[4953]: I1203 00:07:11.375220 4953 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4bb40260-dbaa-4fb0-84df-5e680505d512" path="/var/lib/kubelet/pods/4bb40260-dbaa-4fb0-84df-5e680505d512/volumes" Dec 03 00:07:11 crc kubenswrapper[4953]: I1203 00:07:11.376555 4953 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5225d0e4-402f-4861-b410-819f433b1803" path="/var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes" Dec 03 00:07:11 crc kubenswrapper[4953]: I1203 00:07:11.378388 4953 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5441d097-087c-4d9a-baa8-b210afa90fc9" path="/var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes" Dec 03 00:07:11 crc kubenswrapper[4953]: I1203 00:07:11.379281 4953 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="57a731c4-ef35-47a8-b875-bfb08a7f8011" path="/var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes" Dec 03 00:07:11 crc kubenswrapper[4953]: I1203 00:07:11.380787 4953 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b88f790-22fa-440e-b583-365168c0b23d" path="/var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/volumes" Dec 03 00:07:11 crc kubenswrapper[4953]: I1203 00:07:11.382028 4953 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5fe579f8-e8a6-4643-bce5-a661393c4dde" path="/var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/volumes" Dec 03 00:07:11 crc kubenswrapper[4953]: I1203 00:07:11.383365 4953 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6402fda4-df10-493c-b4e5-d0569419652d" path="/var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes" Dec 03 00:07:11 crc kubenswrapper[4953]: I1203 00:07:11.384181 4953 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6509e943-70c6-444c-bc41-48a544e36fbd" path="/var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes" Dec 03 00:07:11 crc kubenswrapper[4953]: I1203 00:07:11.385404 4953 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6731426b-95fe-49ff-bb5f-40441049fde2" path="/var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/volumes" Dec 03 00:07:11 crc kubenswrapper[4953]: I1203 00:07:11.385568 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:10Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:11Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:11 crc kubenswrapper[4953]: I1203 00:07:11.386897 4953 kubelet_volumes.go:152] "Cleaned up orphaned volume subpath from pod" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volume-subpaths/run-systemd/ovnkube-controller/6" Dec 03 00:07:11 crc kubenswrapper[4953]: I1203 00:07:11.387084 4953 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volumes" Dec 03 00:07:11 crc kubenswrapper[4953]: I1203 00:07:11.388853 4953 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7539238d-5fe0-46ed-884e-1c3b566537ec" path="/var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes" Dec 03 00:07:11 crc kubenswrapper[4953]: I1203 00:07:11.389477 4953 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7583ce53-e0fe-4a16-9e4d-50516596a136" path="/var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes" Dec 03 00:07:11 crc kubenswrapper[4953]: I1203 00:07:11.390005 4953 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7bb08738-c794-4ee8-9972-3a62ca171029" path="/var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes" Dec 03 00:07:11 crc kubenswrapper[4953]: I1203 00:07:11.391902 4953 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87cf06ed-a83f-41a7-828d-70653580a8cb" path="/var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes" Dec 03 00:07:11 crc kubenswrapper[4953]: I1203 00:07:11.394432 4953 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" path="/var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes" Dec 03 00:07:11 crc kubenswrapper[4953]: I1203 00:07:11.395050 4953 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="925f1c65-6136-48ba-85aa-3a3b50560753" path="/var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes" Dec 03 00:07:11 crc kubenswrapper[4953]: I1203 00:07:11.396264 4953 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" path="/var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/volumes" Dec 03 00:07:11 crc kubenswrapper[4953]: I1203 00:07:11.397286 4953 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d4552c7-cd75-42dd-8880-30dd377c49a4" path="/var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes" Dec 03 00:07:11 crc kubenswrapper[4953]: I1203 00:07:11.400930 4953 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" path="/var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/volumes" Dec 03 00:07:11 crc kubenswrapper[4953]: I1203 00:07:11.401726 4953 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a31745f5-9847-4afe-82a5-3161cc66ca93" path="/var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes" Dec 03 00:07:11 crc kubenswrapper[4953]: I1203 00:07:11.405076 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cf821840-5353-4620-8f73-99b56b371ab0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cae5e2994fcc24dc9b2ea73e79572615b3311edc42642b7f278fb84123989b23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://29e4b9d1b70a93fb133e0759ad148ae133766a66efeae3ff06a21dfae0e805ff\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dec43536772aa7c3f5a662fa7f3bd73b1cae2791ad454f85792287828d0ffeed\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://761973393a93c59c23e69f41b8ac0d16abc35abd61ecbc6d51a42d94cd664233\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://761973393a93c59c23e69f41b8ac0d16abc35abd61ecbc6d51a42d94cd664233\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T00:07:09Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 00:07:03.713374 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 00:07:03.715052 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1528563048/tls.crt::/tmp/serving-cert-1528563048/tls.key\\\\\\\"\\\\nI1203 00:07:09.300052 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 00:07:09.304192 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 00:07:09.304255 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 00:07:09.304309 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 00:07:09.304323 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 00:07:09.313563 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 00:07:09.313597 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 00:07:09.313603 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 00:07:09.313609 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 00:07:09.313613 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 00:07:09.313618 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 00:07:09.313622 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 00:07:09.313709 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 00:07:09.327347 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T00:06:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c143d4232fb8bf7ef5999a664e431a00ab0db030f4d31a28d0bb21e00e38581\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:53Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72af6c27296eae2e275d455795c68f0d272869d63642b8353b17d466d0c901cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://72af6c27296eae2e275d455795c68f0d272869d63642b8353b17d466d0c901cf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:06:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:06:51Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:11Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:11 crc kubenswrapper[4953]: I1203 00:07:11.406024 4953 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" path="/var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes" Dec 03 00:07:11 crc kubenswrapper[4953]: I1203 00:07:11.408226 4953 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6312bbd-5731-4ea0-a20f-81d5a57df44a" path="/var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/volumes" Dec 03 00:07:11 crc kubenswrapper[4953]: I1203 00:07:11.409692 4953 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" path="/var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes" Dec 03 00:07:11 crc kubenswrapper[4953]: I1203 00:07:11.410738 4953 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" path="/var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes" Dec 03 00:07:11 crc kubenswrapper[4953]: I1203 00:07:11.412213 4953 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" path="/var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/volumes" Dec 03 00:07:11 crc kubenswrapper[4953]: I1203 00:07:11.413798 4953 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf126b07-da06-4140-9a57-dfd54fc6b486" path="/var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes" Dec 03 00:07:11 crc kubenswrapper[4953]: I1203 00:07:11.415163 4953 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c03ee662-fb2f-4fc4-a2c1-af487c19d254" path="/var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes" Dec 03 00:07:11 crc kubenswrapper[4953]: I1203 00:07:11.415800 4953 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" path="/var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/volumes" Dec 03 00:07:11 crc kubenswrapper[4953]: I1203 00:07:11.416443 4953 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7e6199b-1264-4501-8953-767f51328d08" path="/var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes" Dec 03 00:07:11 crc kubenswrapper[4953]: I1203 00:07:11.417552 4953 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="efdd0498-1daa-4136-9a4a-3b948c2293fc" path="/var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/volumes" Dec 03 00:07:11 crc kubenswrapper[4953]: I1203 00:07:11.418371 4953 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" path="/var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/volumes" Dec 03 00:07:11 crc kubenswrapper[4953]: I1203 00:07:11.420043 4953 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fda69060-fa79-4696-b1a6-7980f124bf7c" path="/var/lib/kubelet/pods/fda69060-fa79-4696-b1a6-7980f124bf7c/volumes" Dec 03 00:07:11 crc kubenswrapper[4953]: I1203 00:07:11.434697 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:10Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:11Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:11 crc kubenswrapper[4953]: I1203 00:07:11.463162 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"e69b9eee8a7df0a76edca28af3234a72b809af642931ce4d821364ca69c5c40a"} Dec 03 00:07:11 crc kubenswrapper[4953]: I1203 00:07:11.465375 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"09657a9f448a4c92a759c73a0a892076d7eb46393e6fd565e5f73f34c71aac23"} Dec 03 00:07:11 crc kubenswrapper[4953]: I1203 00:07:11.465408 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"5af5eac4a58870f63497f99b503cfa5fb922278095e0332ec50376925ecf2109"} Dec 03 00:07:11 crc kubenswrapper[4953]: I1203 00:07:11.465421 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"49772fe1f3c67c7ceb49d288df16e3d1d444c51608d6562659d4e4637116c35c"} Dec 03 00:07:11 crc kubenswrapper[4953]: I1203 00:07:11.466488 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:11Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:11 crc kubenswrapper[4953]: I1203 00:07:11.467889 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"ce8b414e1fc22eb30decb6a6411b672e7f443055cb89ebb029fbf6ffc1d5dca7"} Dec 03 00:07:11 crc kubenswrapper[4953]: I1203 00:07:11.467925 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"668498b97c57a5c7863587b20bb86e4251785fc8a31bc39658045a86caf4f6e9"} Dec 03 00:07:11 crc kubenswrapper[4953]: I1203 00:07:11.470412 4953 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Dec 03 00:07:11 crc kubenswrapper[4953]: I1203 00:07:11.474577 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"034e0bac6892d361700b8f33d6427b8a55e2bd3f89b14ea77a374c6be94d25e2"} Dec 03 00:07:11 crc kubenswrapper[4953]: I1203 00:07:11.474838 4953 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 00:07:11 crc kubenswrapper[4953]: I1203 00:07:11.481552 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:11Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:11 crc kubenswrapper[4953]: I1203 00:07:11.500613 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:10Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:11Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:11 crc kubenswrapper[4953]: I1203 00:07:11.511536 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:10Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:11Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:11 crc kubenswrapper[4953]: I1203 00:07:11.523562 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f94f8c51-c665-4ca3-bf17-7a04ea5c7ade\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72c5cd7dc1c3a8adfedab6ba43a59d80b3efe65f2bb56be95e76cc86a409dcf3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2810ffc1b141fa7da811406053db9d72c708edf5f2bb1dd75e4d52ab35fd24c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c67b4277b20a9f1287f50ef82d0785c9d02543a8e7ed26e0413eae332b866db7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://258d78e1ad9d3dcda6976909ae3fe71aecead5b995fc66b3fd6e2722d80b5ab8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:06:51Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:11Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:11 crc kubenswrapper[4953]: I1203 00:07:11.536090 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cf821840-5353-4620-8f73-99b56b371ab0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cae5e2994fcc24dc9b2ea73e79572615b3311edc42642b7f278fb84123989b23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://29e4b9d1b70a93fb133e0759ad148ae133766a66efeae3ff06a21dfae0e805ff\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dec43536772aa7c3f5a662fa7f3bd73b1cae2791ad454f85792287828d0ffeed\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://034e0bac6892d361700b8f33d6427b8a55e2bd3f89b14ea77a374c6be94d25e2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://761973393a93c59c23e69f41b8ac0d16abc35abd61ecbc6d51a42d94cd664233\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T00:07:09Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 00:07:03.713374 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 00:07:03.715052 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1528563048/tls.crt::/tmp/serving-cert-1528563048/tls.key\\\\\\\"\\\\nI1203 00:07:09.300052 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 00:07:09.304192 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 00:07:09.304255 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 00:07:09.304309 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 00:07:09.304323 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 00:07:09.313563 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 00:07:09.313597 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 00:07:09.313603 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 00:07:09.313609 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 00:07:09.313613 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 00:07:09.313618 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 00:07:09.313622 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 00:07:09.313709 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 00:07:09.327347 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T00:06:53Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c143d4232fb8bf7ef5999a664e431a00ab0db030f4d31a28d0bb21e00e38581\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:53Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72af6c27296eae2e275d455795c68f0d272869d63642b8353b17d466d0c901cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://72af6c27296eae2e275d455795c68f0d272869d63642b8353b17d466d0c901cf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:06:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:06:51Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:11Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:11 crc kubenswrapper[4953]: I1203 00:07:11.551272 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:10Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:11Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:11 crc kubenswrapper[4953]: I1203 00:07:11.565846 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce8b414e1fc22eb30decb6a6411b672e7f443055cb89ebb029fbf6ffc1d5dca7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:11Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:11 crc kubenswrapper[4953]: I1203 00:07:11.581291 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:10Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:11Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:11 crc kubenswrapper[4953]: I1203 00:07:11.595167 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09657a9f448a4c92a759c73a0a892076d7eb46393e6fd565e5f73f34c71aac23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5af5eac4a58870f63497f99b503cfa5fb922278095e0332ec50376925ecf2109\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:11Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:11 crc kubenswrapper[4953]: I1203 00:07:11.609197 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:10Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:11Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:11 crc kubenswrapper[4953]: I1203 00:07:11.627901 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f94f8c51-c665-4ca3-bf17-7a04ea5c7ade\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72c5cd7dc1c3a8adfedab6ba43a59d80b3efe65f2bb56be95e76cc86a409dcf3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2810ffc1b141fa7da811406053db9d72c708edf5f2bb1dd75e4d52ab35fd24c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c67b4277b20a9f1287f50ef82d0785c9d02543a8e7ed26e0413eae332b866db7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://258d78e1ad9d3dcda6976909ae3fe71aecead5b995fc66b3fd6e2722d80b5ab8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:06:51Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:11Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:11 crc kubenswrapper[4953]: I1203 00:07:11.641635 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:11Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:12 crc kubenswrapper[4953]: I1203 00:07:12.014305 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 00:07:12 crc kubenswrapper[4953]: I1203 00:07:12.014385 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 00:07:12 crc kubenswrapper[4953]: I1203 00:07:12.014408 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 00:07:12 crc kubenswrapper[4953]: E1203 00:07:12.014473 4953 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 00:07:14.014442899 +0000 UTC m=+22.998723692 (durationBeforeRetry 2s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 00:07:12 crc kubenswrapper[4953]: E1203 00:07:12.014497 4953 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 03 00:07:12 crc kubenswrapper[4953]: E1203 00:07:12.014512 4953 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 03 00:07:12 crc kubenswrapper[4953]: E1203 00:07:12.014522 4953 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 00:07:12 crc kubenswrapper[4953]: I1203 00:07:12.014587 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 00:07:12 crc kubenswrapper[4953]: E1203 00:07:12.014603 4953 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 03 00:07:12 crc kubenswrapper[4953]: E1203 00:07:12.014640 4953 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-03 00:07:14.014617433 +0000 UTC m=+22.998898286 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 00:07:12 crc kubenswrapper[4953]: E1203 00:07:12.014696 4953 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 03 00:07:12 crc kubenswrapper[4953]: I1203 00:07:12.014701 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 00:07:12 crc kubenswrapper[4953]: E1203 00:07:12.014719 4953 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 03 00:07:12 crc kubenswrapper[4953]: E1203 00:07:12.014719 4953 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-03 00:07:14.014697535 +0000 UTC m=+22.998978358 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 03 00:07:12 crc kubenswrapper[4953]: E1203 00:07:12.014732 4953 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 00:07:12 crc kubenswrapper[4953]: E1203 00:07:12.014776 4953 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-03 00:07:14.014765427 +0000 UTC m=+22.999046220 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 00:07:12 crc kubenswrapper[4953]: E1203 00:07:12.014797 4953 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 03 00:07:12 crc kubenswrapper[4953]: E1203 00:07:12.014838 4953 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-03 00:07:14.014829158 +0000 UTC m=+22.999109951 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 03 00:07:12 crc kubenswrapper[4953]: I1203 00:07:12.347313 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 00:07:12 crc kubenswrapper[4953]: I1203 00:07:12.347393 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 00:07:12 crc kubenswrapper[4953]: I1203 00:07:12.347407 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 00:07:12 crc kubenswrapper[4953]: E1203 00:07:12.347519 4953 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 00:07:12 crc kubenswrapper[4953]: E1203 00:07:12.347829 4953 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 00:07:12 crc kubenswrapper[4953]: E1203 00:07:12.347722 4953 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 00:07:12 crc kubenswrapper[4953]: I1203 00:07:12.488009 4953 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 00:07:12 crc kubenswrapper[4953]: I1203 00:07:12.489683 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:12 crc kubenswrapper[4953]: I1203 00:07:12.489735 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:12 crc kubenswrapper[4953]: I1203 00:07:12.489754 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:12 crc kubenswrapper[4953]: I1203 00:07:12.489832 4953 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 03 00:07:12 crc kubenswrapper[4953]: I1203 00:07:12.500138 4953 kubelet_node_status.go:115] "Node was previously registered" node="crc" Dec 03 00:07:12 crc kubenswrapper[4953]: I1203 00:07:12.500462 4953 kubelet_node_status.go:79] "Successfully registered node" node="crc" Dec 03 00:07:12 crc kubenswrapper[4953]: I1203 00:07:12.501608 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:12 crc kubenswrapper[4953]: I1203 00:07:12.501661 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:12 crc kubenswrapper[4953]: I1203 00:07:12.501675 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:12 crc kubenswrapper[4953]: I1203 00:07:12.501694 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:12 crc kubenswrapper[4953]: I1203 00:07:12.501706 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:12Z","lastTransitionTime":"2025-12-03T00:07:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:12 crc kubenswrapper[4953]: E1203 00:07:12.530774 4953 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:07:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:12Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:07:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:12Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:07:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:12Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:07:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:12Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"eb4ae5a6-0f6f-49ef-a027-1cdee9d3c536\\\",\\\"systemUUID\\\":\\\"4c1f3d44-3d3d-425d-993f-471223206a89\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:12Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:12 crc kubenswrapper[4953]: I1203 00:07:12.534218 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:12 crc kubenswrapper[4953]: I1203 00:07:12.534280 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:12 crc kubenswrapper[4953]: I1203 00:07:12.534299 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:12 crc kubenswrapper[4953]: I1203 00:07:12.534321 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:12 crc kubenswrapper[4953]: I1203 00:07:12.534339 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:12Z","lastTransitionTime":"2025-12-03T00:07:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:12 crc kubenswrapper[4953]: E1203 00:07:12.549108 4953 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:07:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:12Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:07:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:12Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:07:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:12Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:07:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:12Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"eb4ae5a6-0f6f-49ef-a027-1cdee9d3c536\\\",\\\"systemUUID\\\":\\\"4c1f3d44-3d3d-425d-993f-471223206a89\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:12Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:12 crc kubenswrapper[4953]: I1203 00:07:12.554648 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:12 crc kubenswrapper[4953]: I1203 00:07:12.554765 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:12 crc kubenswrapper[4953]: I1203 00:07:12.554793 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:12 crc kubenswrapper[4953]: I1203 00:07:12.555024 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:12 crc kubenswrapper[4953]: I1203 00:07:12.555122 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:12Z","lastTransitionTime":"2025-12-03T00:07:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:12 crc kubenswrapper[4953]: E1203 00:07:12.573318 4953 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:07:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:12Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:07:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:12Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:07:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:12Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:07:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:12Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"eb4ae5a6-0f6f-49ef-a027-1cdee9d3c536\\\",\\\"systemUUID\\\":\\\"4c1f3d44-3d3d-425d-993f-471223206a89\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:12Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:12 crc kubenswrapper[4953]: I1203 00:07:12.578885 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:12 crc kubenswrapper[4953]: I1203 00:07:12.578960 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:12 crc kubenswrapper[4953]: I1203 00:07:12.579008 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:12 crc kubenswrapper[4953]: I1203 00:07:12.579029 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:12 crc kubenswrapper[4953]: I1203 00:07:12.579043 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:12Z","lastTransitionTime":"2025-12-03T00:07:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:12 crc kubenswrapper[4953]: E1203 00:07:12.600704 4953 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:07:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:12Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:07:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:12Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:07:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:12Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:07:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:12Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"eb4ae5a6-0f6f-49ef-a027-1cdee9d3c536\\\",\\\"systemUUID\\\":\\\"4c1f3d44-3d3d-425d-993f-471223206a89\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:12Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:12 crc kubenswrapper[4953]: I1203 00:07:12.604640 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:12 crc kubenswrapper[4953]: I1203 00:07:12.604705 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:12 crc kubenswrapper[4953]: I1203 00:07:12.604718 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:12 crc kubenswrapper[4953]: I1203 00:07:12.604735 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:12 crc kubenswrapper[4953]: I1203 00:07:12.604746 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:12Z","lastTransitionTime":"2025-12-03T00:07:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:12 crc kubenswrapper[4953]: E1203 00:07:12.617236 4953 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:07:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:12Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:07:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:12Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:07:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:12Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:07:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:12Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"eb4ae5a6-0f6f-49ef-a027-1cdee9d3c536\\\",\\\"systemUUID\\\":\\\"4c1f3d44-3d3d-425d-993f-471223206a89\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:12Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:12 crc kubenswrapper[4953]: E1203 00:07:12.617392 4953 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 03 00:07:12 crc kubenswrapper[4953]: I1203 00:07:12.618949 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:12 crc kubenswrapper[4953]: I1203 00:07:12.618995 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:12 crc kubenswrapper[4953]: I1203 00:07:12.619007 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:12 crc kubenswrapper[4953]: I1203 00:07:12.619024 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:12 crc kubenswrapper[4953]: I1203 00:07:12.619036 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:12Z","lastTransitionTime":"2025-12-03T00:07:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:12 crc kubenswrapper[4953]: I1203 00:07:12.722051 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:12 crc kubenswrapper[4953]: I1203 00:07:12.722090 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:12 crc kubenswrapper[4953]: I1203 00:07:12.722102 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:12 crc kubenswrapper[4953]: I1203 00:07:12.722118 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:12 crc kubenswrapper[4953]: I1203 00:07:12.722130 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:12Z","lastTransitionTime":"2025-12-03T00:07:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:12 crc kubenswrapper[4953]: I1203 00:07:12.824512 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:12 crc kubenswrapper[4953]: I1203 00:07:12.824584 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:12 crc kubenswrapper[4953]: I1203 00:07:12.824602 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:12 crc kubenswrapper[4953]: I1203 00:07:12.824626 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:12 crc kubenswrapper[4953]: I1203 00:07:12.824645 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:12Z","lastTransitionTime":"2025-12-03T00:07:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:12 crc kubenswrapper[4953]: I1203 00:07:12.928143 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:12 crc kubenswrapper[4953]: I1203 00:07:12.928202 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:12 crc kubenswrapper[4953]: I1203 00:07:12.928213 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:12 crc kubenswrapper[4953]: I1203 00:07:12.928239 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:12 crc kubenswrapper[4953]: I1203 00:07:12.928254 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:12Z","lastTransitionTime":"2025-12-03T00:07:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:13 crc kubenswrapper[4953]: I1203 00:07:13.031077 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:13 crc kubenswrapper[4953]: I1203 00:07:13.031160 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:13 crc kubenswrapper[4953]: I1203 00:07:13.031174 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:13 crc kubenswrapper[4953]: I1203 00:07:13.031199 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:13 crc kubenswrapper[4953]: I1203 00:07:13.031215 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:13Z","lastTransitionTime":"2025-12-03T00:07:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:13 crc kubenswrapper[4953]: I1203 00:07:13.134282 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:13 crc kubenswrapper[4953]: I1203 00:07:13.134352 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:13 crc kubenswrapper[4953]: I1203 00:07:13.134373 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:13 crc kubenswrapper[4953]: I1203 00:07:13.134399 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:13 crc kubenswrapper[4953]: I1203 00:07:13.134420 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:13Z","lastTransitionTime":"2025-12-03T00:07:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:13 crc kubenswrapper[4953]: I1203 00:07:13.245383 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:13 crc kubenswrapper[4953]: I1203 00:07:13.245437 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:13 crc kubenswrapper[4953]: I1203 00:07:13.245449 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:13 crc kubenswrapper[4953]: I1203 00:07:13.245471 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:13 crc kubenswrapper[4953]: I1203 00:07:13.245487 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:13Z","lastTransitionTime":"2025-12-03T00:07:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:13 crc kubenswrapper[4953]: I1203 00:07:13.349415 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:13 crc kubenswrapper[4953]: I1203 00:07:13.349505 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:13 crc kubenswrapper[4953]: I1203 00:07:13.349524 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:13 crc kubenswrapper[4953]: I1203 00:07:13.349553 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:13 crc kubenswrapper[4953]: I1203 00:07:13.349568 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:13Z","lastTransitionTime":"2025-12-03T00:07:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:13 crc kubenswrapper[4953]: I1203 00:07:13.452154 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:13 crc kubenswrapper[4953]: I1203 00:07:13.452225 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:13 crc kubenswrapper[4953]: I1203 00:07:13.452237 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:13 crc kubenswrapper[4953]: I1203 00:07:13.452256 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:13 crc kubenswrapper[4953]: I1203 00:07:13.452268 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:13Z","lastTransitionTime":"2025-12-03T00:07:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:13 crc kubenswrapper[4953]: I1203 00:07:13.481699 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"f2ed9c7fa96c8951fbc1728fe7a91fefdd999077285621df9c1abedf9ae0c449"} Dec 03 00:07:13 crc kubenswrapper[4953]: I1203 00:07:13.498652 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cf821840-5353-4620-8f73-99b56b371ab0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cae5e2994fcc24dc9b2ea73e79572615b3311edc42642b7f278fb84123989b23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://29e4b9d1b70a93fb133e0759ad148ae133766a66efeae3ff06a21dfae0e805ff\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dec43536772aa7c3f5a662fa7f3bd73b1cae2791ad454f85792287828d0ffeed\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://034e0bac6892d361700b8f33d6427b8a55e2bd3f89b14ea77a374c6be94d25e2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://761973393a93c59c23e69f41b8ac0d16abc35abd61ecbc6d51a42d94cd664233\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T00:07:09Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 00:07:03.713374 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 00:07:03.715052 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1528563048/tls.crt::/tmp/serving-cert-1528563048/tls.key\\\\\\\"\\\\nI1203 00:07:09.300052 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 00:07:09.304192 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 00:07:09.304255 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 00:07:09.304309 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 00:07:09.304323 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 00:07:09.313563 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 00:07:09.313597 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 00:07:09.313603 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 00:07:09.313609 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 00:07:09.313613 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 00:07:09.313618 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 00:07:09.313622 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 00:07:09.313709 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 00:07:09.327347 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T00:06:53Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c143d4232fb8bf7ef5999a664e431a00ab0db030f4d31a28d0bb21e00e38581\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:53Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72af6c27296eae2e275d455795c68f0d272869d63642b8353b17d466d0c901cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://72af6c27296eae2e275d455795c68f0d272869d63642b8353b17d466d0c901cf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:06:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:06:51Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:13Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:13 crc kubenswrapper[4953]: I1203 00:07:13.518106 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:10Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:13Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:13 crc kubenswrapper[4953]: I1203 00:07:13.535258 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce8b414e1fc22eb30decb6a6411b672e7f443055cb89ebb029fbf6ffc1d5dca7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:13Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:13 crc kubenswrapper[4953]: I1203 00:07:13.548964 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:10Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:13Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:13 crc kubenswrapper[4953]: I1203 00:07:13.554335 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:13 crc kubenswrapper[4953]: I1203 00:07:13.554406 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:13 crc kubenswrapper[4953]: I1203 00:07:13.554560 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:13 crc kubenswrapper[4953]: I1203 00:07:13.554857 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:13 crc kubenswrapper[4953]: I1203 00:07:13.554879 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:13Z","lastTransitionTime":"2025-12-03T00:07:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:13 crc kubenswrapper[4953]: I1203 00:07:13.566215 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f94f8c51-c665-4ca3-bf17-7a04ea5c7ade\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72c5cd7dc1c3a8adfedab6ba43a59d80b3efe65f2bb56be95e76cc86a409dcf3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2810ffc1b141fa7da811406053db9d72c708edf5f2bb1dd75e4d52ab35fd24c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c67b4277b20a9f1287f50ef82d0785c9d02543a8e7ed26e0413eae332b866db7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://258d78e1ad9d3dcda6976909ae3fe71aecead5b995fc66b3fd6e2722d80b5ab8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:06:51Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:13Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:13 crc kubenswrapper[4953]: I1203 00:07:13.580225 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:13Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:13 crc kubenswrapper[4953]: I1203 00:07:13.597154 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09657a9f448a4c92a759c73a0a892076d7eb46393e6fd565e5f73f34c71aac23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5af5eac4a58870f63497f99b503cfa5fb922278095e0332ec50376925ecf2109\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:13Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:13 crc kubenswrapper[4953]: I1203 00:07:13.613183 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2ed9c7fa96c8951fbc1728fe7a91fefdd999077285621df9c1abedf9ae0c449\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:13Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:13 crc kubenswrapper[4953]: I1203 00:07:13.657614 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:13 crc kubenswrapper[4953]: I1203 00:07:13.657666 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:13 crc kubenswrapper[4953]: I1203 00:07:13.657684 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:13 crc kubenswrapper[4953]: I1203 00:07:13.657708 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:13 crc kubenswrapper[4953]: I1203 00:07:13.657726 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:13Z","lastTransitionTime":"2025-12-03T00:07:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:13 crc kubenswrapper[4953]: I1203 00:07:13.760269 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:13 crc kubenswrapper[4953]: I1203 00:07:13.760308 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:13 crc kubenswrapper[4953]: I1203 00:07:13.760319 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:13 crc kubenswrapper[4953]: I1203 00:07:13.760333 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:13 crc kubenswrapper[4953]: I1203 00:07:13.760342 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:13Z","lastTransitionTime":"2025-12-03T00:07:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:13 crc kubenswrapper[4953]: I1203 00:07:13.862909 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:13 crc kubenswrapper[4953]: I1203 00:07:13.863015 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:13 crc kubenswrapper[4953]: I1203 00:07:13.863034 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:13 crc kubenswrapper[4953]: I1203 00:07:13.863060 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:13 crc kubenswrapper[4953]: I1203 00:07:13.863078 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:13Z","lastTransitionTime":"2025-12-03T00:07:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:13 crc kubenswrapper[4953]: I1203 00:07:13.965496 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:13 crc kubenswrapper[4953]: I1203 00:07:13.965534 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:13 crc kubenswrapper[4953]: I1203 00:07:13.965544 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:13 crc kubenswrapper[4953]: I1203 00:07:13.965559 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:13 crc kubenswrapper[4953]: I1203 00:07:13.965568 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:13Z","lastTransitionTime":"2025-12-03T00:07:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:14 crc kubenswrapper[4953]: I1203 00:07:14.032387 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 00:07:14 crc kubenswrapper[4953]: I1203 00:07:14.032484 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 00:07:14 crc kubenswrapper[4953]: I1203 00:07:14.032514 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 00:07:14 crc kubenswrapper[4953]: I1203 00:07:14.032537 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 00:07:14 crc kubenswrapper[4953]: I1203 00:07:14.032562 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 00:07:14 crc kubenswrapper[4953]: E1203 00:07:14.032651 4953 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 03 00:07:14 crc kubenswrapper[4953]: E1203 00:07:14.032700 4953 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 00:07:18.032662355 +0000 UTC m=+27.016943168 (durationBeforeRetry 4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 00:07:14 crc kubenswrapper[4953]: E1203 00:07:14.032772 4953 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-03 00:07:18.032759247 +0000 UTC m=+27.017040060 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 03 00:07:14 crc kubenswrapper[4953]: E1203 00:07:14.032886 4953 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 03 00:07:14 crc kubenswrapper[4953]: E1203 00:07:14.032921 4953 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 03 00:07:14 crc kubenswrapper[4953]: E1203 00:07:14.032945 4953 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 03 00:07:14 crc kubenswrapper[4953]: E1203 00:07:14.032967 4953 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 03 00:07:14 crc kubenswrapper[4953]: E1203 00:07:14.033010 4953 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 00:07:14 crc kubenswrapper[4953]: E1203 00:07:14.033034 4953 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 03 00:07:14 crc kubenswrapper[4953]: E1203 00:07:14.033055 4953 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 00:07:14 crc kubenswrapper[4953]: E1203 00:07:14.033060 4953 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-03 00:07:18.033039034 +0000 UTC m=+27.017319917 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 03 00:07:14 crc kubenswrapper[4953]: E1203 00:07:14.033128 4953 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-03 00:07:18.033089675 +0000 UTC m=+27.017370528 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 00:07:14 crc kubenswrapper[4953]: E1203 00:07:14.033168 4953 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-03 00:07:18.033150657 +0000 UTC m=+27.017431620 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 00:07:14 crc kubenswrapper[4953]: I1203 00:07:14.068263 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:14 crc kubenswrapper[4953]: I1203 00:07:14.068314 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:14 crc kubenswrapper[4953]: I1203 00:07:14.068329 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:14 crc kubenswrapper[4953]: I1203 00:07:14.068351 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:14 crc kubenswrapper[4953]: I1203 00:07:14.068368 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:14Z","lastTransitionTime":"2025-12-03T00:07:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:14 crc kubenswrapper[4953]: I1203 00:07:14.171425 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:14 crc kubenswrapper[4953]: I1203 00:07:14.171513 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:14 crc kubenswrapper[4953]: I1203 00:07:14.171530 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:14 crc kubenswrapper[4953]: I1203 00:07:14.171557 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:14 crc kubenswrapper[4953]: I1203 00:07:14.171577 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:14Z","lastTransitionTime":"2025-12-03T00:07:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:14 crc kubenswrapper[4953]: I1203 00:07:14.275706 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:14 crc kubenswrapper[4953]: I1203 00:07:14.275848 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:14 crc kubenswrapper[4953]: I1203 00:07:14.275870 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:14 crc kubenswrapper[4953]: I1203 00:07:14.275894 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:14 crc kubenswrapper[4953]: I1203 00:07:14.275912 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:14Z","lastTransitionTime":"2025-12-03T00:07:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:14 crc kubenswrapper[4953]: I1203 00:07:14.348119 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 00:07:14 crc kubenswrapper[4953]: I1203 00:07:14.348155 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 00:07:14 crc kubenswrapper[4953]: I1203 00:07:14.348132 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 00:07:14 crc kubenswrapper[4953]: E1203 00:07:14.348340 4953 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 00:07:14 crc kubenswrapper[4953]: E1203 00:07:14.348461 4953 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 00:07:14 crc kubenswrapper[4953]: E1203 00:07:14.348561 4953 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 00:07:14 crc kubenswrapper[4953]: I1203 00:07:14.378738 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:14 crc kubenswrapper[4953]: I1203 00:07:14.378796 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:14 crc kubenswrapper[4953]: I1203 00:07:14.378818 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:14 crc kubenswrapper[4953]: I1203 00:07:14.378845 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:14 crc kubenswrapper[4953]: I1203 00:07:14.378865 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:14Z","lastTransitionTime":"2025-12-03T00:07:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:14 crc kubenswrapper[4953]: I1203 00:07:14.482028 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:14 crc kubenswrapper[4953]: I1203 00:07:14.482115 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:14 crc kubenswrapper[4953]: I1203 00:07:14.482141 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:14 crc kubenswrapper[4953]: I1203 00:07:14.482173 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:14 crc kubenswrapper[4953]: I1203 00:07:14.482197 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:14Z","lastTransitionTime":"2025-12-03T00:07:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:14 crc kubenswrapper[4953]: I1203 00:07:14.585880 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:14 crc kubenswrapper[4953]: I1203 00:07:14.585921 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:14 crc kubenswrapper[4953]: I1203 00:07:14.585937 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:14 crc kubenswrapper[4953]: I1203 00:07:14.585960 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:14 crc kubenswrapper[4953]: I1203 00:07:14.586011 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:14Z","lastTransitionTime":"2025-12-03T00:07:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:14 crc kubenswrapper[4953]: I1203 00:07:14.688895 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:14 crc kubenswrapper[4953]: I1203 00:07:14.689063 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:14 crc kubenswrapper[4953]: I1203 00:07:14.689091 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:14 crc kubenswrapper[4953]: I1203 00:07:14.689119 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:14 crc kubenswrapper[4953]: I1203 00:07:14.689138 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:14Z","lastTransitionTime":"2025-12-03T00:07:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:14 crc kubenswrapper[4953]: I1203 00:07:14.791562 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:14 crc kubenswrapper[4953]: I1203 00:07:14.791672 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:14 crc kubenswrapper[4953]: I1203 00:07:14.791692 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:14 crc kubenswrapper[4953]: I1203 00:07:14.791716 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:14 crc kubenswrapper[4953]: I1203 00:07:14.791737 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:14Z","lastTransitionTime":"2025-12-03T00:07:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:14 crc kubenswrapper[4953]: I1203 00:07:14.895355 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:14 crc kubenswrapper[4953]: I1203 00:07:14.895428 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:14 crc kubenswrapper[4953]: I1203 00:07:14.895452 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:14 crc kubenswrapper[4953]: I1203 00:07:14.895482 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:14 crc kubenswrapper[4953]: I1203 00:07:14.895504 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:14Z","lastTransitionTime":"2025-12-03T00:07:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:14 crc kubenswrapper[4953]: I1203 00:07:14.998019 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:14 crc kubenswrapper[4953]: I1203 00:07:14.998055 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:14 crc kubenswrapper[4953]: I1203 00:07:14.998067 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:14 crc kubenswrapper[4953]: I1203 00:07:14.998082 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:14 crc kubenswrapper[4953]: I1203 00:07:14.998092 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:14Z","lastTransitionTime":"2025-12-03T00:07:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:15 crc kubenswrapper[4953]: I1203 00:07:15.100285 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:15 crc kubenswrapper[4953]: I1203 00:07:15.100342 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:15 crc kubenswrapper[4953]: I1203 00:07:15.100360 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:15 crc kubenswrapper[4953]: I1203 00:07:15.100383 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:15 crc kubenswrapper[4953]: I1203 00:07:15.100400 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:15Z","lastTransitionTime":"2025-12-03T00:07:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:15 crc kubenswrapper[4953]: I1203 00:07:15.203135 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:15 crc kubenswrapper[4953]: I1203 00:07:15.203195 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:15 crc kubenswrapper[4953]: I1203 00:07:15.203224 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:15 crc kubenswrapper[4953]: I1203 00:07:15.203252 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:15 crc kubenswrapper[4953]: I1203 00:07:15.203273 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:15Z","lastTransitionTime":"2025-12-03T00:07:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:15 crc kubenswrapper[4953]: I1203 00:07:15.305872 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:15 crc kubenswrapper[4953]: I1203 00:07:15.306058 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:15 crc kubenswrapper[4953]: I1203 00:07:15.306089 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:15 crc kubenswrapper[4953]: I1203 00:07:15.306117 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:15 crc kubenswrapper[4953]: I1203 00:07:15.306139 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:15Z","lastTransitionTime":"2025-12-03T00:07:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:15 crc kubenswrapper[4953]: I1203 00:07:15.409666 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:15 crc kubenswrapper[4953]: I1203 00:07:15.409719 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:15 crc kubenswrapper[4953]: I1203 00:07:15.409732 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:15 crc kubenswrapper[4953]: I1203 00:07:15.409750 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:15 crc kubenswrapper[4953]: I1203 00:07:15.409762 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:15Z","lastTransitionTime":"2025-12-03T00:07:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:15 crc kubenswrapper[4953]: I1203 00:07:15.512552 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:15 crc kubenswrapper[4953]: I1203 00:07:15.512592 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:15 crc kubenswrapper[4953]: I1203 00:07:15.512603 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:15 crc kubenswrapper[4953]: I1203 00:07:15.512620 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:15 crc kubenswrapper[4953]: I1203 00:07:15.512631 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:15Z","lastTransitionTime":"2025-12-03T00:07:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:15 crc kubenswrapper[4953]: I1203 00:07:15.615138 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:15 crc kubenswrapper[4953]: I1203 00:07:15.615206 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:15 crc kubenswrapper[4953]: I1203 00:07:15.615219 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:15 crc kubenswrapper[4953]: I1203 00:07:15.615241 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:15 crc kubenswrapper[4953]: I1203 00:07:15.615253 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:15Z","lastTransitionTime":"2025-12-03T00:07:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:15 crc kubenswrapper[4953]: I1203 00:07:15.717303 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:15 crc kubenswrapper[4953]: I1203 00:07:15.717338 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:15 crc kubenswrapper[4953]: I1203 00:07:15.717349 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:15 crc kubenswrapper[4953]: I1203 00:07:15.717364 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:15 crc kubenswrapper[4953]: I1203 00:07:15.717391 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:15Z","lastTransitionTime":"2025-12-03T00:07:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:15 crc kubenswrapper[4953]: I1203 00:07:15.819459 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:15 crc kubenswrapper[4953]: I1203 00:07:15.819487 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:15 crc kubenswrapper[4953]: I1203 00:07:15.819495 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:15 crc kubenswrapper[4953]: I1203 00:07:15.819512 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:15 crc kubenswrapper[4953]: I1203 00:07:15.819529 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:15Z","lastTransitionTime":"2025-12-03T00:07:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:15 crc kubenswrapper[4953]: I1203 00:07:15.922180 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:15 crc kubenswrapper[4953]: I1203 00:07:15.922219 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:15 crc kubenswrapper[4953]: I1203 00:07:15.922230 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:15 crc kubenswrapper[4953]: I1203 00:07:15.922256 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:15 crc kubenswrapper[4953]: I1203 00:07:15.922267 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:15Z","lastTransitionTime":"2025-12-03T00:07:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:16 crc kubenswrapper[4953]: I1203 00:07:16.024850 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:16 crc kubenswrapper[4953]: I1203 00:07:16.024898 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:16 crc kubenswrapper[4953]: I1203 00:07:16.024909 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:16 crc kubenswrapper[4953]: I1203 00:07:16.024925 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:16 crc kubenswrapper[4953]: I1203 00:07:16.024936 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:16Z","lastTransitionTime":"2025-12-03T00:07:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:16 crc kubenswrapper[4953]: I1203 00:07:16.127414 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:16 crc kubenswrapper[4953]: I1203 00:07:16.127462 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:16 crc kubenswrapper[4953]: I1203 00:07:16.127474 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:16 crc kubenswrapper[4953]: I1203 00:07:16.127494 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:16 crc kubenswrapper[4953]: I1203 00:07:16.127506 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:16Z","lastTransitionTime":"2025-12-03T00:07:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:16 crc kubenswrapper[4953]: I1203 00:07:16.210517 4953 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-675qk"] Dec 03 00:07:16 crc kubenswrapper[4953]: I1203 00:07:16.211746 4953 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-additional-cni-plugins-wj6rz"] Dec 03 00:07:16 crc kubenswrapper[4953]: I1203 00:07:16.213277 4953 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/node-resolver-kl5s9"] Dec 03 00:07:16 crc kubenswrapper[4953]: I1203 00:07:16.214839 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-675qk" Dec 03 00:07:16 crc kubenswrapper[4953]: I1203 00:07:16.215085 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-wj6rz" Dec 03 00:07:16 crc kubenswrapper[4953]: I1203 00:07:16.216465 4953 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-daemon-6dfbn"] Dec 03 00:07:16 crc kubenswrapper[4953]: I1203 00:07:16.216701 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-kl5s9" Dec 03 00:07:16 crc kubenswrapper[4953]: I1203 00:07:16.219409 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Dec 03 00:07:16 crc kubenswrapper[4953]: I1203 00:07:16.219642 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Dec 03 00:07:16 crc kubenswrapper[4953]: I1203 00:07:16.219832 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Dec 03 00:07:16 crc kubenswrapper[4953]: I1203 00:07:16.219960 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-6dfbn" Dec 03 00:07:16 crc kubenswrapper[4953]: I1203 00:07:16.223553 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Dec 03 00:07:16 crc kubenswrapper[4953]: I1203 00:07:16.232642 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Dec 03 00:07:16 crc kubenswrapper[4953]: I1203 00:07:16.232719 4953 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Dec 03 00:07:16 crc kubenswrapper[4953]: I1203 00:07:16.232727 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Dec 03 00:07:16 crc kubenswrapper[4953]: I1203 00:07:16.232838 4953 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Dec 03 00:07:16 crc kubenswrapper[4953]: I1203 00:07:16.232887 4953 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Dec 03 00:07:16 crc kubenswrapper[4953]: I1203 00:07:16.233102 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:16 crc kubenswrapper[4953]: I1203 00:07:16.233135 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:16 crc kubenswrapper[4953]: I1203 00:07:16.233145 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:16 crc kubenswrapper[4953]: I1203 00:07:16.233162 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:16 crc kubenswrapper[4953]: I1203 00:07:16.233173 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:16Z","lastTransitionTime":"2025-12-03T00:07:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:16 crc kubenswrapper[4953]: I1203 00:07:16.234552 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Dec 03 00:07:16 crc kubenswrapper[4953]: I1203 00:07:16.234561 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Dec 03 00:07:16 crc kubenswrapper[4953]: I1203 00:07:16.237360 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Dec 03 00:07:16 crc kubenswrapper[4953]: I1203 00:07:16.237752 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Dec 03 00:07:16 crc kubenswrapper[4953]: I1203 00:07:16.237955 4953 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Dec 03 00:07:16 crc kubenswrapper[4953]: I1203 00:07:16.238166 4953 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Dec 03 00:07:16 crc kubenswrapper[4953]: I1203 00:07:16.253217 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/96aa67a0-15dd-420d-a1d5-ed239d7df1fb-system-cni-dir\") pod \"multus-additional-cni-plugins-wj6rz\" (UID: \"96aa67a0-15dd-420d-a1d5-ed239d7df1fb\") " pod="openshift-multus/multus-additional-cni-plugins-wj6rz" Dec 03 00:07:16 crc kubenswrapper[4953]: I1203 00:07:16.253280 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/43a4bad8-71f4-441a-8302-aeb526a986af-host-var-lib-cni-multus\") pod \"multus-675qk\" (UID: \"43a4bad8-71f4-441a-8302-aeb526a986af\") " pod="openshift-multus/multus-675qk" Dec 03 00:07:16 crc kubenswrapper[4953]: I1203 00:07:16.253307 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kf62m\" (UniqueName: \"kubernetes.io/projected/96aa67a0-15dd-420d-a1d5-ed239d7df1fb-kube-api-access-kf62m\") pod \"multus-additional-cni-plugins-wj6rz\" (UID: \"96aa67a0-15dd-420d-a1d5-ed239d7df1fb\") " pod="openshift-multus/multus-additional-cni-plugins-wj6rz" Dec 03 00:07:16 crc kubenswrapper[4953]: I1203 00:07:16.253331 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/43a4bad8-71f4-441a-8302-aeb526a986af-multus-daemon-config\") pod \"multus-675qk\" (UID: \"43a4bad8-71f4-441a-8302-aeb526a986af\") " pod="openshift-multus/multus-675qk" Dec 03 00:07:16 crc kubenswrapper[4953]: I1203 00:07:16.253353 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/43a4bad8-71f4-441a-8302-aeb526a986af-system-cni-dir\") pod \"multus-675qk\" (UID: \"43a4bad8-71f4-441a-8302-aeb526a986af\") " pod="openshift-multus/multus-675qk" Dec 03 00:07:16 crc kubenswrapper[4953]: I1203 00:07:16.253428 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ct26c\" (UniqueName: \"kubernetes.io/projected/53ffb1e9-9a2e-4869-80d7-776f4aa7ef8b-kube-api-access-ct26c\") pod \"machine-config-daemon-6dfbn\" (UID: \"53ffb1e9-9a2e-4869-80d7-776f4aa7ef8b\") " pod="openshift-machine-config-operator/machine-config-daemon-6dfbn" Dec 03 00:07:16 crc kubenswrapper[4953]: I1203 00:07:16.253501 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/43a4bad8-71f4-441a-8302-aeb526a986af-host-run-multus-certs\") pod \"multus-675qk\" (UID: \"43a4bad8-71f4-441a-8302-aeb526a986af\") " pod="openshift-multus/multus-675qk" Dec 03 00:07:16 crc kubenswrapper[4953]: I1203 00:07:16.253528 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/43a4bad8-71f4-441a-8302-aeb526a986af-host-run-netns\") pod \"multus-675qk\" (UID: \"43a4bad8-71f4-441a-8302-aeb526a986af\") " pod="openshift-multus/multus-675qk" Dec 03 00:07:16 crc kubenswrapper[4953]: I1203 00:07:16.253560 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/d85e5c97-16ad-4bd7-9e73-b7f12678be97-hosts-file\") pod \"node-resolver-kl5s9\" (UID: \"d85e5c97-16ad-4bd7-9e73-b7f12678be97\") " pod="openshift-dns/node-resolver-kl5s9" Dec 03 00:07:16 crc kubenswrapper[4953]: I1203 00:07:16.253607 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gdrw4\" (UniqueName: \"kubernetes.io/projected/43a4bad8-71f4-441a-8302-aeb526a986af-kube-api-access-gdrw4\") pod \"multus-675qk\" (UID: \"43a4bad8-71f4-441a-8302-aeb526a986af\") " pod="openshift-multus/multus-675qk" Dec 03 00:07:16 crc kubenswrapper[4953]: I1203 00:07:16.253628 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j24lb\" (UniqueName: \"kubernetes.io/projected/d85e5c97-16ad-4bd7-9e73-b7f12678be97-kube-api-access-j24lb\") pod \"node-resolver-kl5s9\" (UID: \"d85e5c97-16ad-4bd7-9e73-b7f12678be97\") " pod="openshift-dns/node-resolver-kl5s9" Dec 03 00:07:16 crc kubenswrapper[4953]: I1203 00:07:16.253657 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/43a4bad8-71f4-441a-8302-aeb526a986af-host-var-lib-cni-bin\") pod \"multus-675qk\" (UID: \"43a4bad8-71f4-441a-8302-aeb526a986af\") " pod="openshift-multus/multus-675qk" Dec 03 00:07:16 crc kubenswrapper[4953]: I1203 00:07:16.253679 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/53ffb1e9-9a2e-4869-80d7-776f4aa7ef8b-mcd-auth-proxy-config\") pod \"machine-config-daemon-6dfbn\" (UID: \"53ffb1e9-9a2e-4869-80d7-776f4aa7ef8b\") " pod="openshift-machine-config-operator/machine-config-daemon-6dfbn" Dec 03 00:07:16 crc kubenswrapper[4953]: I1203 00:07:16.253747 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/96aa67a0-15dd-420d-a1d5-ed239d7df1fb-tuning-conf-dir\") pod \"multus-additional-cni-plugins-wj6rz\" (UID: \"96aa67a0-15dd-420d-a1d5-ed239d7df1fb\") " pod="openshift-multus/multus-additional-cni-plugins-wj6rz" Dec 03 00:07:16 crc kubenswrapper[4953]: I1203 00:07:16.253780 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/43a4bad8-71f4-441a-8302-aeb526a986af-cnibin\") pod \"multus-675qk\" (UID: \"43a4bad8-71f4-441a-8302-aeb526a986af\") " pod="openshift-multus/multus-675qk" Dec 03 00:07:16 crc kubenswrapper[4953]: I1203 00:07:16.253805 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/43a4bad8-71f4-441a-8302-aeb526a986af-host-var-lib-kubelet\") pod \"multus-675qk\" (UID: \"43a4bad8-71f4-441a-8302-aeb526a986af\") " pod="openshift-multus/multus-675qk" Dec 03 00:07:16 crc kubenswrapper[4953]: I1203 00:07:16.253827 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/43a4bad8-71f4-441a-8302-aeb526a986af-multus-conf-dir\") pod \"multus-675qk\" (UID: \"43a4bad8-71f4-441a-8302-aeb526a986af\") " pod="openshift-multus/multus-675qk" Dec 03 00:07:16 crc kubenswrapper[4953]: I1203 00:07:16.253847 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/43a4bad8-71f4-441a-8302-aeb526a986af-os-release\") pod \"multus-675qk\" (UID: \"43a4bad8-71f4-441a-8302-aeb526a986af\") " pod="openshift-multus/multus-675qk" Dec 03 00:07:16 crc kubenswrapper[4953]: I1203 00:07:16.253871 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/43a4bad8-71f4-441a-8302-aeb526a986af-host-run-k8s-cni-cncf-io\") pod \"multus-675qk\" (UID: \"43a4bad8-71f4-441a-8302-aeb526a986af\") " pod="openshift-multus/multus-675qk" Dec 03 00:07:16 crc kubenswrapper[4953]: I1203 00:07:16.253891 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/43a4bad8-71f4-441a-8302-aeb526a986af-etc-kubernetes\") pod \"multus-675qk\" (UID: \"43a4bad8-71f4-441a-8302-aeb526a986af\") " pod="openshift-multus/multus-675qk" Dec 03 00:07:16 crc kubenswrapper[4953]: I1203 00:07:16.253919 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/96aa67a0-15dd-420d-a1d5-ed239d7df1fb-cnibin\") pod \"multus-additional-cni-plugins-wj6rz\" (UID: \"96aa67a0-15dd-420d-a1d5-ed239d7df1fb\") " pod="openshift-multus/multus-additional-cni-plugins-wj6rz" Dec 03 00:07:16 crc kubenswrapper[4953]: I1203 00:07:16.253938 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/43a4bad8-71f4-441a-8302-aeb526a986af-multus-cni-dir\") pod \"multus-675qk\" (UID: \"43a4bad8-71f4-441a-8302-aeb526a986af\") " pod="openshift-multus/multus-675qk" Dec 03 00:07:16 crc kubenswrapper[4953]: I1203 00:07:16.253990 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/53ffb1e9-9a2e-4869-80d7-776f4aa7ef8b-proxy-tls\") pod \"machine-config-daemon-6dfbn\" (UID: \"53ffb1e9-9a2e-4869-80d7-776f4aa7ef8b\") " pod="openshift-machine-config-operator/machine-config-daemon-6dfbn" Dec 03 00:07:16 crc kubenswrapper[4953]: I1203 00:07:16.254019 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/96aa67a0-15dd-420d-a1d5-ed239d7df1fb-cni-binary-copy\") pod \"multus-additional-cni-plugins-wj6rz\" (UID: \"96aa67a0-15dd-420d-a1d5-ed239d7df1fb\") " pod="openshift-multus/multus-additional-cni-plugins-wj6rz" Dec 03 00:07:16 crc kubenswrapper[4953]: I1203 00:07:16.254044 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/96aa67a0-15dd-420d-a1d5-ed239d7df1fb-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-wj6rz\" (UID: \"96aa67a0-15dd-420d-a1d5-ed239d7df1fb\") " pod="openshift-multus/multus-additional-cni-plugins-wj6rz" Dec 03 00:07:16 crc kubenswrapper[4953]: I1203 00:07:16.254075 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/43a4bad8-71f4-441a-8302-aeb526a986af-hostroot\") pod \"multus-675qk\" (UID: \"43a4bad8-71f4-441a-8302-aeb526a986af\") " pod="openshift-multus/multus-675qk" Dec 03 00:07:16 crc kubenswrapper[4953]: I1203 00:07:16.254101 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/43a4bad8-71f4-441a-8302-aeb526a986af-cni-binary-copy\") pod \"multus-675qk\" (UID: \"43a4bad8-71f4-441a-8302-aeb526a986af\") " pod="openshift-multus/multus-675qk" Dec 03 00:07:16 crc kubenswrapper[4953]: I1203 00:07:16.254136 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/53ffb1e9-9a2e-4869-80d7-776f4aa7ef8b-rootfs\") pod \"machine-config-daemon-6dfbn\" (UID: \"53ffb1e9-9a2e-4869-80d7-776f4aa7ef8b\") " pod="openshift-machine-config-operator/machine-config-daemon-6dfbn" Dec 03 00:07:16 crc kubenswrapper[4953]: I1203 00:07:16.254158 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/96aa67a0-15dd-420d-a1d5-ed239d7df1fb-os-release\") pod \"multus-additional-cni-plugins-wj6rz\" (UID: \"96aa67a0-15dd-420d-a1d5-ed239d7df1fb\") " pod="openshift-multus/multus-additional-cni-plugins-wj6rz" Dec 03 00:07:16 crc kubenswrapper[4953]: I1203 00:07:16.254179 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/43a4bad8-71f4-441a-8302-aeb526a986af-multus-socket-dir-parent\") pod \"multus-675qk\" (UID: \"43a4bad8-71f4-441a-8302-aeb526a986af\") " pod="openshift-multus/multus-675qk" Dec 03 00:07:16 crc kubenswrapper[4953]: I1203 00:07:16.260050 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f94f8c51-c665-4ca3-bf17-7a04ea5c7ade\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72c5cd7dc1c3a8adfedab6ba43a59d80b3efe65f2bb56be95e76cc86a409dcf3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2810ffc1b141fa7da811406053db9d72c708edf5f2bb1dd75e4d52ab35fd24c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c67b4277b20a9f1287f50ef82d0785c9d02543a8e7ed26e0413eae332b866db7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://258d78e1ad9d3dcda6976909ae3fe71aecead5b995fc66b3fd6e2722d80b5ab8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:06:51Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:16Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:16 crc kubenswrapper[4953]: I1203 00:07:16.296430 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:16Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:16 crc kubenswrapper[4953]: I1203 00:07:16.324805 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09657a9f448a4c92a759c73a0a892076d7eb46393e6fd565e5f73f34c71aac23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5af5eac4a58870f63497f99b503cfa5fb922278095e0332ec50376925ecf2109\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:16Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:16 crc kubenswrapper[4953]: I1203 00:07:16.336018 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:16 crc kubenswrapper[4953]: I1203 00:07:16.336062 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:16 crc kubenswrapper[4953]: I1203 00:07:16.336075 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:16 crc kubenswrapper[4953]: I1203 00:07:16.336096 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:16 crc kubenswrapper[4953]: I1203 00:07:16.336110 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:16Z","lastTransitionTime":"2025-12-03T00:07:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:16 crc kubenswrapper[4953]: I1203 00:07:16.338496 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2ed9c7fa96c8951fbc1728fe7a91fefdd999077285621df9c1abedf9ae0c449\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:16Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:16 crc kubenswrapper[4953]: I1203 00:07:16.347200 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 00:07:16 crc kubenswrapper[4953]: I1203 00:07:16.347241 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 00:07:16 crc kubenswrapper[4953]: I1203 00:07:16.347199 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 00:07:16 crc kubenswrapper[4953]: E1203 00:07:16.347341 4953 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 00:07:16 crc kubenswrapper[4953]: E1203 00:07:16.347459 4953 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 00:07:16 crc kubenswrapper[4953]: E1203 00:07:16.347531 4953 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 00:07:16 crc kubenswrapper[4953]: I1203 00:07:16.350002 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-wj6rz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"96aa67a0-15dd-420d-a1d5-ed239d7df1fb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:16Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:16Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:16Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kf62m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kf62m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kf62m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kf62m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kf62m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kf62m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kf62m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:07:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-wj6rz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:16Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:16 crc kubenswrapper[4953]: I1203 00:07:16.354934 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/53ffb1e9-9a2e-4869-80d7-776f4aa7ef8b-proxy-tls\") pod \"machine-config-daemon-6dfbn\" (UID: \"53ffb1e9-9a2e-4869-80d7-776f4aa7ef8b\") " pod="openshift-machine-config-operator/machine-config-daemon-6dfbn" Dec 03 00:07:16 crc kubenswrapper[4953]: I1203 00:07:16.354994 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/96aa67a0-15dd-420d-a1d5-ed239d7df1fb-cni-binary-copy\") pod \"multus-additional-cni-plugins-wj6rz\" (UID: \"96aa67a0-15dd-420d-a1d5-ed239d7df1fb\") " pod="openshift-multus/multus-additional-cni-plugins-wj6rz" Dec 03 00:07:16 crc kubenswrapper[4953]: I1203 00:07:16.355015 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/96aa67a0-15dd-420d-a1d5-ed239d7df1fb-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-wj6rz\" (UID: \"96aa67a0-15dd-420d-a1d5-ed239d7df1fb\") " pod="openshift-multus/multus-additional-cni-plugins-wj6rz" Dec 03 00:07:16 crc kubenswrapper[4953]: I1203 00:07:16.355036 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/43a4bad8-71f4-441a-8302-aeb526a986af-hostroot\") pod \"multus-675qk\" (UID: \"43a4bad8-71f4-441a-8302-aeb526a986af\") " pod="openshift-multus/multus-675qk" Dec 03 00:07:16 crc kubenswrapper[4953]: I1203 00:07:16.355053 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/43a4bad8-71f4-441a-8302-aeb526a986af-cni-binary-copy\") pod \"multus-675qk\" (UID: \"43a4bad8-71f4-441a-8302-aeb526a986af\") " pod="openshift-multus/multus-675qk" Dec 03 00:07:16 crc kubenswrapper[4953]: I1203 00:07:16.355085 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/53ffb1e9-9a2e-4869-80d7-776f4aa7ef8b-rootfs\") pod \"machine-config-daemon-6dfbn\" (UID: \"53ffb1e9-9a2e-4869-80d7-776f4aa7ef8b\") " pod="openshift-machine-config-operator/machine-config-daemon-6dfbn" Dec 03 00:07:16 crc kubenswrapper[4953]: I1203 00:07:16.355102 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/43a4bad8-71f4-441a-8302-aeb526a986af-multus-socket-dir-parent\") pod \"multus-675qk\" (UID: \"43a4bad8-71f4-441a-8302-aeb526a986af\") " pod="openshift-multus/multus-675qk" Dec 03 00:07:16 crc kubenswrapper[4953]: I1203 00:07:16.355117 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/96aa67a0-15dd-420d-a1d5-ed239d7df1fb-os-release\") pod \"multus-additional-cni-plugins-wj6rz\" (UID: \"96aa67a0-15dd-420d-a1d5-ed239d7df1fb\") " pod="openshift-multus/multus-additional-cni-plugins-wj6rz" Dec 03 00:07:16 crc kubenswrapper[4953]: I1203 00:07:16.355184 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/43a4bad8-71f4-441a-8302-aeb526a986af-hostroot\") pod \"multus-675qk\" (UID: \"43a4bad8-71f4-441a-8302-aeb526a986af\") " pod="openshift-multus/multus-675qk" Dec 03 00:07:16 crc kubenswrapper[4953]: I1203 00:07:16.355221 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/53ffb1e9-9a2e-4869-80d7-776f4aa7ef8b-rootfs\") pod \"machine-config-daemon-6dfbn\" (UID: \"53ffb1e9-9a2e-4869-80d7-776f4aa7ef8b\") " pod="openshift-machine-config-operator/machine-config-daemon-6dfbn" Dec 03 00:07:16 crc kubenswrapper[4953]: I1203 00:07:16.355264 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/43a4bad8-71f4-441a-8302-aeb526a986af-multus-socket-dir-parent\") pod \"multus-675qk\" (UID: \"43a4bad8-71f4-441a-8302-aeb526a986af\") " pod="openshift-multus/multus-675qk" Dec 03 00:07:16 crc kubenswrapper[4953]: I1203 00:07:16.355285 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/96aa67a0-15dd-420d-a1d5-ed239d7df1fb-os-release\") pod \"multus-additional-cni-plugins-wj6rz\" (UID: \"96aa67a0-15dd-420d-a1d5-ed239d7df1fb\") " pod="openshift-multus/multus-additional-cni-plugins-wj6rz" Dec 03 00:07:16 crc kubenswrapper[4953]: I1203 00:07:16.355364 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/96aa67a0-15dd-420d-a1d5-ed239d7df1fb-system-cni-dir\") pod \"multus-additional-cni-plugins-wj6rz\" (UID: \"96aa67a0-15dd-420d-a1d5-ed239d7df1fb\") " pod="openshift-multus/multus-additional-cni-plugins-wj6rz" Dec 03 00:07:16 crc kubenswrapper[4953]: I1203 00:07:16.355548 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/96aa67a0-15dd-420d-a1d5-ed239d7df1fb-system-cni-dir\") pod \"multus-additional-cni-plugins-wj6rz\" (UID: \"96aa67a0-15dd-420d-a1d5-ed239d7df1fb\") " pod="openshift-multus/multus-additional-cni-plugins-wj6rz" Dec 03 00:07:16 crc kubenswrapper[4953]: I1203 00:07:16.355582 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/43a4bad8-71f4-441a-8302-aeb526a986af-host-var-lib-cni-multus\") pod \"multus-675qk\" (UID: \"43a4bad8-71f4-441a-8302-aeb526a986af\") " pod="openshift-multus/multus-675qk" Dec 03 00:07:16 crc kubenswrapper[4953]: I1203 00:07:16.355645 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/43a4bad8-71f4-441a-8302-aeb526a986af-host-var-lib-cni-multus\") pod \"multus-675qk\" (UID: \"43a4bad8-71f4-441a-8302-aeb526a986af\") " pod="openshift-multus/multus-675qk" Dec 03 00:07:16 crc kubenswrapper[4953]: I1203 00:07:16.355716 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kf62m\" (UniqueName: \"kubernetes.io/projected/96aa67a0-15dd-420d-a1d5-ed239d7df1fb-kube-api-access-kf62m\") pod \"multus-additional-cni-plugins-wj6rz\" (UID: \"96aa67a0-15dd-420d-a1d5-ed239d7df1fb\") " pod="openshift-multus/multus-additional-cni-plugins-wj6rz" Dec 03 00:07:16 crc kubenswrapper[4953]: I1203 00:07:16.355733 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/43a4bad8-71f4-441a-8302-aeb526a986af-multus-daemon-config\") pod \"multus-675qk\" (UID: \"43a4bad8-71f4-441a-8302-aeb526a986af\") " pod="openshift-multus/multus-675qk" Dec 03 00:07:16 crc kubenswrapper[4953]: I1203 00:07:16.355756 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/96aa67a0-15dd-420d-a1d5-ed239d7df1fb-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-wj6rz\" (UID: \"96aa67a0-15dd-420d-a1d5-ed239d7df1fb\") " pod="openshift-multus/multus-additional-cni-plugins-wj6rz" Dec 03 00:07:16 crc kubenswrapper[4953]: I1203 00:07:16.355777 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/96aa67a0-15dd-420d-a1d5-ed239d7df1fb-cni-binary-copy\") pod \"multus-additional-cni-plugins-wj6rz\" (UID: \"96aa67a0-15dd-420d-a1d5-ed239d7df1fb\") " pod="openshift-multus/multus-additional-cni-plugins-wj6rz" Dec 03 00:07:16 crc kubenswrapper[4953]: I1203 00:07:16.355874 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/43a4bad8-71f4-441a-8302-aeb526a986af-cni-binary-copy\") pod \"multus-675qk\" (UID: \"43a4bad8-71f4-441a-8302-aeb526a986af\") " pod="openshift-multus/multus-675qk" Dec 03 00:07:16 crc kubenswrapper[4953]: I1203 00:07:16.356278 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ct26c\" (UniqueName: \"kubernetes.io/projected/53ffb1e9-9a2e-4869-80d7-776f4aa7ef8b-kube-api-access-ct26c\") pod \"machine-config-daemon-6dfbn\" (UID: \"53ffb1e9-9a2e-4869-80d7-776f4aa7ef8b\") " pod="openshift-machine-config-operator/machine-config-daemon-6dfbn" Dec 03 00:07:16 crc kubenswrapper[4953]: I1203 00:07:16.356378 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/43a4bad8-71f4-441a-8302-aeb526a986af-multus-daemon-config\") pod \"multus-675qk\" (UID: \"43a4bad8-71f4-441a-8302-aeb526a986af\") " pod="openshift-multus/multus-675qk" Dec 03 00:07:16 crc kubenswrapper[4953]: I1203 00:07:16.356511 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/43a4bad8-71f4-441a-8302-aeb526a986af-system-cni-dir\") pod \"multus-675qk\" (UID: \"43a4bad8-71f4-441a-8302-aeb526a986af\") " pod="openshift-multus/multus-675qk" Dec 03 00:07:16 crc kubenswrapper[4953]: I1203 00:07:16.356680 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/43a4bad8-71f4-441a-8302-aeb526a986af-system-cni-dir\") pod \"multus-675qk\" (UID: \"43a4bad8-71f4-441a-8302-aeb526a986af\") " pod="openshift-multus/multus-675qk" Dec 03 00:07:16 crc kubenswrapper[4953]: I1203 00:07:16.356790 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/43a4bad8-71f4-441a-8302-aeb526a986af-host-run-multus-certs\") pod \"multus-675qk\" (UID: \"43a4bad8-71f4-441a-8302-aeb526a986af\") " pod="openshift-multus/multus-675qk" Dec 03 00:07:16 crc kubenswrapper[4953]: I1203 00:07:16.356960 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/43a4bad8-71f4-441a-8302-aeb526a986af-host-run-netns\") pod \"multus-675qk\" (UID: \"43a4bad8-71f4-441a-8302-aeb526a986af\") " pod="openshift-multus/multus-675qk" Dec 03 00:07:16 crc kubenswrapper[4953]: I1203 00:07:16.357130 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/d85e5c97-16ad-4bd7-9e73-b7f12678be97-hosts-file\") pod \"node-resolver-kl5s9\" (UID: \"d85e5c97-16ad-4bd7-9e73-b7f12678be97\") " pod="openshift-dns/node-resolver-kl5s9" Dec 03 00:07:16 crc kubenswrapper[4953]: I1203 00:07:16.357300 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gdrw4\" (UniqueName: \"kubernetes.io/projected/43a4bad8-71f4-441a-8302-aeb526a986af-kube-api-access-gdrw4\") pod \"multus-675qk\" (UID: \"43a4bad8-71f4-441a-8302-aeb526a986af\") " pod="openshift-multus/multus-675qk" Dec 03 00:07:16 crc kubenswrapper[4953]: I1203 00:07:16.357409 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j24lb\" (UniqueName: \"kubernetes.io/projected/d85e5c97-16ad-4bd7-9e73-b7f12678be97-kube-api-access-j24lb\") pod \"node-resolver-kl5s9\" (UID: \"d85e5c97-16ad-4bd7-9e73-b7f12678be97\") " pod="openshift-dns/node-resolver-kl5s9" Dec 03 00:07:16 crc kubenswrapper[4953]: I1203 00:07:16.357507 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/43a4bad8-71f4-441a-8302-aeb526a986af-host-var-lib-cni-bin\") pod \"multus-675qk\" (UID: \"43a4bad8-71f4-441a-8302-aeb526a986af\") " pod="openshift-multus/multus-675qk" Dec 03 00:07:16 crc kubenswrapper[4953]: I1203 00:07:16.357704 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/53ffb1e9-9a2e-4869-80d7-776f4aa7ef8b-mcd-auth-proxy-config\") pod \"machine-config-daemon-6dfbn\" (UID: \"53ffb1e9-9a2e-4869-80d7-776f4aa7ef8b\") " pod="openshift-machine-config-operator/machine-config-daemon-6dfbn" Dec 03 00:07:16 crc kubenswrapper[4953]: I1203 00:07:16.357809 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/43a4bad8-71f4-441a-8302-aeb526a986af-host-var-lib-kubelet\") pod \"multus-675qk\" (UID: \"43a4bad8-71f4-441a-8302-aeb526a986af\") " pod="openshift-multus/multus-675qk" Dec 03 00:07:16 crc kubenswrapper[4953]: I1203 00:07:16.357906 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/43a4bad8-71f4-441a-8302-aeb526a986af-multus-conf-dir\") pod \"multus-675qk\" (UID: \"43a4bad8-71f4-441a-8302-aeb526a986af\") " pod="openshift-multus/multus-675qk" Dec 03 00:07:16 crc kubenswrapper[4953]: I1203 00:07:16.358024 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/96aa67a0-15dd-420d-a1d5-ed239d7df1fb-tuning-conf-dir\") pod \"multus-additional-cni-plugins-wj6rz\" (UID: \"96aa67a0-15dd-420d-a1d5-ed239d7df1fb\") " pod="openshift-multus/multus-additional-cni-plugins-wj6rz" Dec 03 00:07:16 crc kubenswrapper[4953]: I1203 00:07:16.358133 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/43a4bad8-71f4-441a-8302-aeb526a986af-cnibin\") pod \"multus-675qk\" (UID: \"43a4bad8-71f4-441a-8302-aeb526a986af\") " pod="openshift-multus/multus-675qk" Dec 03 00:07:16 crc kubenswrapper[4953]: I1203 00:07:16.357083 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/43a4bad8-71f4-441a-8302-aeb526a986af-host-run-netns\") pod \"multus-675qk\" (UID: \"43a4bad8-71f4-441a-8302-aeb526a986af\") " pod="openshift-multus/multus-675qk" Dec 03 00:07:16 crc kubenswrapper[4953]: I1203 00:07:16.358324 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/43a4bad8-71f4-441a-8302-aeb526a986af-os-release\") pod \"multus-675qk\" (UID: \"43a4bad8-71f4-441a-8302-aeb526a986af\") " pod="openshift-multus/multus-675qk" Dec 03 00:07:16 crc kubenswrapper[4953]: I1203 00:07:16.358416 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/43a4bad8-71f4-441a-8302-aeb526a986af-host-var-lib-kubelet\") pod \"multus-675qk\" (UID: \"43a4bad8-71f4-441a-8302-aeb526a986af\") " pod="openshift-multus/multus-675qk" Dec 03 00:07:16 crc kubenswrapper[4953]: I1203 00:07:16.358425 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/43a4bad8-71f4-441a-8302-aeb526a986af-host-run-k8s-cni-cncf-io\") pod \"multus-675qk\" (UID: \"43a4bad8-71f4-441a-8302-aeb526a986af\") " pod="openshift-multus/multus-675qk" Dec 03 00:07:16 crc kubenswrapper[4953]: I1203 00:07:16.358478 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/43a4bad8-71f4-441a-8302-aeb526a986af-etc-kubernetes\") pod \"multus-675qk\" (UID: \"43a4bad8-71f4-441a-8302-aeb526a986af\") " pod="openshift-multus/multus-675qk" Dec 03 00:07:16 crc kubenswrapper[4953]: I1203 00:07:16.358503 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/96aa67a0-15dd-420d-a1d5-ed239d7df1fb-cnibin\") pod \"multus-additional-cni-plugins-wj6rz\" (UID: \"96aa67a0-15dd-420d-a1d5-ed239d7df1fb\") " pod="openshift-multus/multus-additional-cni-plugins-wj6rz" Dec 03 00:07:16 crc kubenswrapper[4953]: I1203 00:07:16.358523 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/43a4bad8-71f4-441a-8302-aeb526a986af-multus-cni-dir\") pod \"multus-675qk\" (UID: \"43a4bad8-71f4-441a-8302-aeb526a986af\") " pod="openshift-multus/multus-675qk" Dec 03 00:07:16 crc kubenswrapper[4953]: I1203 00:07:16.358627 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/43a4bad8-71f4-441a-8302-aeb526a986af-multus-cni-dir\") pod \"multus-675qk\" (UID: \"43a4bad8-71f4-441a-8302-aeb526a986af\") " pod="openshift-multus/multus-675qk" Dec 03 00:07:16 crc kubenswrapper[4953]: I1203 00:07:16.357267 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/d85e5c97-16ad-4bd7-9e73-b7f12678be97-hosts-file\") pod \"node-resolver-kl5s9\" (UID: \"d85e5c97-16ad-4bd7-9e73-b7f12678be97\") " pod="openshift-dns/node-resolver-kl5s9" Dec 03 00:07:16 crc kubenswrapper[4953]: I1203 00:07:16.358667 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/43a4bad8-71f4-441a-8302-aeb526a986af-host-var-lib-cni-bin\") pod \"multus-675qk\" (UID: \"43a4bad8-71f4-441a-8302-aeb526a986af\") " pod="openshift-multus/multus-675qk" Dec 03 00:07:16 crc kubenswrapper[4953]: I1203 00:07:16.358825 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/43a4bad8-71f4-441a-8302-aeb526a986af-host-run-k8s-cni-cncf-io\") pod \"multus-675qk\" (UID: \"43a4bad8-71f4-441a-8302-aeb526a986af\") " pod="openshift-multus/multus-675qk" Dec 03 00:07:16 crc kubenswrapper[4953]: I1203 00:07:16.358893 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/43a4bad8-71f4-441a-8302-aeb526a986af-cnibin\") pod \"multus-675qk\" (UID: \"43a4bad8-71f4-441a-8302-aeb526a986af\") " pod="openshift-multus/multus-675qk" Dec 03 00:07:16 crc kubenswrapper[4953]: I1203 00:07:16.358916 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/43a4bad8-71f4-441a-8302-aeb526a986af-multus-conf-dir\") pod \"multus-675qk\" (UID: \"43a4bad8-71f4-441a-8302-aeb526a986af\") " pod="openshift-multus/multus-675qk" Dec 03 00:07:16 crc kubenswrapper[4953]: I1203 00:07:16.358954 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/43a4bad8-71f4-441a-8302-aeb526a986af-etc-kubernetes\") pod \"multus-675qk\" (UID: \"43a4bad8-71f4-441a-8302-aeb526a986af\") " pod="openshift-multus/multus-675qk" Dec 03 00:07:16 crc kubenswrapper[4953]: I1203 00:07:16.358997 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/96aa67a0-15dd-420d-a1d5-ed239d7df1fb-cnibin\") pod \"multus-additional-cni-plugins-wj6rz\" (UID: \"96aa67a0-15dd-420d-a1d5-ed239d7df1fb\") " pod="openshift-multus/multus-additional-cni-plugins-wj6rz" Dec 03 00:07:16 crc kubenswrapper[4953]: I1203 00:07:16.356922 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/43a4bad8-71f4-441a-8302-aeb526a986af-host-run-multus-certs\") pod \"multus-675qk\" (UID: \"43a4bad8-71f4-441a-8302-aeb526a986af\") " pod="openshift-multus/multus-675qk" Dec 03 00:07:16 crc kubenswrapper[4953]: I1203 00:07:16.359014 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/43a4bad8-71f4-441a-8302-aeb526a986af-os-release\") pod \"multus-675qk\" (UID: \"43a4bad8-71f4-441a-8302-aeb526a986af\") " pod="openshift-multus/multus-675qk" Dec 03 00:07:16 crc kubenswrapper[4953]: I1203 00:07:16.359286 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/96aa67a0-15dd-420d-a1d5-ed239d7df1fb-tuning-conf-dir\") pod \"multus-additional-cni-plugins-wj6rz\" (UID: \"96aa67a0-15dd-420d-a1d5-ed239d7df1fb\") " pod="openshift-multus/multus-additional-cni-plugins-wj6rz" Dec 03 00:07:16 crc kubenswrapper[4953]: I1203 00:07:16.359315 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/53ffb1e9-9a2e-4869-80d7-776f4aa7ef8b-mcd-auth-proxy-config\") pod \"machine-config-daemon-6dfbn\" (UID: \"53ffb1e9-9a2e-4869-80d7-776f4aa7ef8b\") " pod="openshift-machine-config-operator/machine-config-daemon-6dfbn" Dec 03 00:07:16 crc kubenswrapper[4953]: I1203 00:07:16.363551 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce8b414e1fc22eb30decb6a6411b672e7f443055cb89ebb029fbf6ffc1d5dca7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:16Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:16 crc kubenswrapper[4953]: I1203 00:07:16.367555 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/53ffb1e9-9a2e-4869-80d7-776f4aa7ef8b-proxy-tls\") pod \"machine-config-daemon-6dfbn\" (UID: \"53ffb1e9-9a2e-4869-80d7-776f4aa7ef8b\") " pod="openshift-machine-config-operator/machine-config-daemon-6dfbn" Dec 03 00:07:16 crc kubenswrapper[4953]: I1203 00:07:16.372997 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gdrw4\" (UniqueName: \"kubernetes.io/projected/43a4bad8-71f4-441a-8302-aeb526a986af-kube-api-access-gdrw4\") pod \"multus-675qk\" (UID: \"43a4bad8-71f4-441a-8302-aeb526a986af\") " pod="openshift-multus/multus-675qk" Dec 03 00:07:16 crc kubenswrapper[4953]: I1203 00:07:16.374161 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j24lb\" (UniqueName: \"kubernetes.io/projected/d85e5c97-16ad-4bd7-9e73-b7f12678be97-kube-api-access-j24lb\") pod \"node-resolver-kl5s9\" (UID: \"d85e5c97-16ad-4bd7-9e73-b7f12678be97\") " pod="openshift-dns/node-resolver-kl5s9" Dec 03 00:07:16 crc kubenswrapper[4953]: I1203 00:07:16.375154 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ct26c\" (UniqueName: \"kubernetes.io/projected/53ffb1e9-9a2e-4869-80d7-776f4aa7ef8b-kube-api-access-ct26c\") pod \"machine-config-daemon-6dfbn\" (UID: \"53ffb1e9-9a2e-4869-80d7-776f4aa7ef8b\") " pod="openshift-machine-config-operator/machine-config-daemon-6dfbn" Dec 03 00:07:16 crc kubenswrapper[4953]: I1203 00:07:16.378110 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:10Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:16Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:16 crc kubenswrapper[4953]: I1203 00:07:16.379804 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kf62m\" (UniqueName: \"kubernetes.io/projected/96aa67a0-15dd-420d-a1d5-ed239d7df1fb-kube-api-access-kf62m\") pod \"multus-additional-cni-plugins-wj6rz\" (UID: \"96aa67a0-15dd-420d-a1d5-ed239d7df1fb\") " pod="openshift-multus/multus-additional-cni-plugins-wj6rz" Dec 03 00:07:16 crc kubenswrapper[4953]: I1203 00:07:16.390400 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-675qk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"43a4bad8-71f4-441a-8302-aeb526a986af\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:16Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:16Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gdrw4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:07:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-675qk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:16Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:16 crc kubenswrapper[4953]: I1203 00:07:16.403244 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cf821840-5353-4620-8f73-99b56b371ab0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cae5e2994fcc24dc9b2ea73e79572615b3311edc42642b7f278fb84123989b23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://29e4b9d1b70a93fb133e0759ad148ae133766a66efeae3ff06a21dfae0e805ff\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dec43536772aa7c3f5a662fa7f3bd73b1cae2791ad454f85792287828d0ffeed\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://034e0bac6892d361700b8f33d6427b8a55e2bd3f89b14ea77a374c6be94d25e2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://761973393a93c59c23e69f41b8ac0d16abc35abd61ecbc6d51a42d94cd664233\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T00:07:09Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 00:07:03.713374 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 00:07:03.715052 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1528563048/tls.crt::/tmp/serving-cert-1528563048/tls.key\\\\\\\"\\\\nI1203 00:07:09.300052 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 00:07:09.304192 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 00:07:09.304255 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 00:07:09.304309 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 00:07:09.304323 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 00:07:09.313563 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 00:07:09.313597 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 00:07:09.313603 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 00:07:09.313609 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 00:07:09.313613 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 00:07:09.313618 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 00:07:09.313622 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 00:07:09.313709 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 00:07:09.327347 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T00:06:53Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c143d4232fb8bf7ef5999a664e431a00ab0db030f4d31a28d0bb21e00e38581\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:53Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72af6c27296eae2e275d455795c68f0d272869d63642b8353b17d466d0c901cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://72af6c27296eae2e275d455795c68f0d272869d63642b8353b17d466d0c901cf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:06:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:06:51Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:16Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:16 crc kubenswrapper[4953]: I1203 00:07:16.418641 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:10Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:16Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:16 crc kubenswrapper[4953]: I1203 00:07:16.434679 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:16Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:16 crc kubenswrapper[4953]: I1203 00:07:16.438220 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:16 crc kubenswrapper[4953]: I1203 00:07:16.438255 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:16 crc kubenswrapper[4953]: I1203 00:07:16.438267 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:16 crc kubenswrapper[4953]: I1203 00:07:16.438285 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:16 crc kubenswrapper[4953]: I1203 00:07:16.438299 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:16Z","lastTransitionTime":"2025-12-03T00:07:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:16 crc kubenswrapper[4953]: I1203 00:07:16.448019 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2ed9c7fa96c8951fbc1728fe7a91fefdd999077285621df9c1abedf9ae0c449\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:16Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:16 crc kubenswrapper[4953]: I1203 00:07:16.461343 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:10Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:16Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:16 crc kubenswrapper[4953]: I1203 00:07:16.477348 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-675qk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"43a4bad8-71f4-441a-8302-aeb526a986af\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:16Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:16Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gdrw4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:07:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-675qk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:16Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:16 crc kubenswrapper[4953]: I1203 00:07:16.491481 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f94f8c51-c665-4ca3-bf17-7a04ea5c7ade\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72c5cd7dc1c3a8adfedab6ba43a59d80b3efe65f2bb56be95e76cc86a409dcf3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2810ffc1b141fa7da811406053db9d72c708edf5f2bb1dd75e4d52ab35fd24c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c67b4277b20a9f1287f50ef82d0785c9d02543a8e7ed26e0413eae332b866db7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://258d78e1ad9d3dcda6976909ae3fe71aecead5b995fc66b3fd6e2722d80b5ab8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:06:51Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:16Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:16 crc kubenswrapper[4953]: I1203 00:07:16.504069 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09657a9f448a4c92a759c73a0a892076d7eb46393e6fd565e5f73f34c71aac23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5af5eac4a58870f63497f99b503cfa5fb922278095e0332ec50376925ecf2109\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:16Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:16 crc kubenswrapper[4953]: I1203 00:07:16.524453 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-wj6rz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"96aa67a0-15dd-420d-a1d5-ed239d7df1fb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:16Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:16Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:16Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kf62m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kf62m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kf62m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kf62m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kf62m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kf62m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kf62m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:07:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-wj6rz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:16Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:16 crc kubenswrapper[4953]: I1203 00:07:16.537854 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-kl5s9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d85e5c97-16ad-4bd7-9e73-b7f12678be97\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:16Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:16Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j24lb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:07:16Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-kl5s9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:16Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:16 crc kubenswrapper[4953]: I1203 00:07:16.538197 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-675qk" Dec 03 00:07:16 crc kubenswrapper[4953]: I1203 00:07:16.540176 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:16 crc kubenswrapper[4953]: I1203 00:07:16.540205 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:16 crc kubenswrapper[4953]: I1203 00:07:16.540214 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:16 crc kubenswrapper[4953]: I1203 00:07:16.540229 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:16 crc kubenswrapper[4953]: I1203 00:07:16.540239 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:16Z","lastTransitionTime":"2025-12-03T00:07:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:16 crc kubenswrapper[4953]: I1203 00:07:16.541378 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-wj6rz" Dec 03 00:07:16 crc kubenswrapper[4953]: I1203 00:07:16.549138 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-kl5s9" Dec 03 00:07:16 crc kubenswrapper[4953]: W1203 00:07:16.552245 4953 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod43a4bad8_71f4_441a_8302_aeb526a986af.slice/crio-7de0dff7d485526cf2e1a3c09a3fe9326b95f15c567b0fb543a4cd54ddbaca95 WatchSource:0}: Error finding container 7de0dff7d485526cf2e1a3c09a3fe9326b95f15c567b0fb543a4cd54ddbaca95: Status 404 returned error can't find the container with id 7de0dff7d485526cf2e1a3c09a3fe9326b95f15c567b0fb543a4cd54ddbaca95 Dec 03 00:07:16 crc kubenswrapper[4953]: I1203 00:07:16.555716 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cf821840-5353-4620-8f73-99b56b371ab0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cae5e2994fcc24dc9b2ea73e79572615b3311edc42642b7f278fb84123989b23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://29e4b9d1b70a93fb133e0759ad148ae133766a66efeae3ff06a21dfae0e805ff\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dec43536772aa7c3f5a662fa7f3bd73b1cae2791ad454f85792287828d0ffeed\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://034e0bac6892d361700b8f33d6427b8a55e2bd3f89b14ea77a374c6be94d25e2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://761973393a93c59c23e69f41b8ac0d16abc35abd61ecbc6d51a42d94cd664233\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T00:07:09Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 00:07:03.713374 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 00:07:03.715052 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1528563048/tls.crt::/tmp/serving-cert-1528563048/tls.key\\\\\\\"\\\\nI1203 00:07:09.300052 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 00:07:09.304192 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 00:07:09.304255 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 00:07:09.304309 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 00:07:09.304323 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 00:07:09.313563 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 00:07:09.313597 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 00:07:09.313603 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 00:07:09.313609 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 00:07:09.313613 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 00:07:09.313618 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 00:07:09.313622 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 00:07:09.313709 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 00:07:09.327347 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T00:06:53Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c143d4232fb8bf7ef5999a664e431a00ab0db030f4d31a28d0bb21e00e38581\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:53Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72af6c27296eae2e275d455795c68f0d272869d63642b8353b17d466d0c901cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://72af6c27296eae2e275d455795c68f0d272869d63642b8353b17d466d0c901cf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:06:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:06:51Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:16Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:16 crc kubenswrapper[4953]: I1203 00:07:16.555959 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-6dfbn" Dec 03 00:07:16 crc kubenswrapper[4953]: W1203 00:07:16.560344 4953 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod96aa67a0_15dd_420d_a1d5_ed239d7df1fb.slice/crio-1e55d024f1dfd02c91fab473e8d5808cf6546138cedd3a2191b5cdb50782047e WatchSource:0}: Error finding container 1e55d024f1dfd02c91fab473e8d5808cf6546138cedd3a2191b5cdb50782047e: Status 404 returned error can't find the container with id 1e55d024f1dfd02c91fab473e8d5808cf6546138cedd3a2191b5cdb50782047e Dec 03 00:07:16 crc kubenswrapper[4953]: I1203 00:07:16.571407 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:10Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:16Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:16 crc kubenswrapper[4953]: W1203 00:07:16.571697 4953 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod53ffb1e9_9a2e_4869_80d7_776f4aa7ef8b.slice/crio-0b0e03de635e73a91cf5f667c42079d951f513a5f6184502cf203a5c683199c8 WatchSource:0}: Error finding container 0b0e03de635e73a91cf5f667c42079d951f513a5f6184502cf203a5c683199c8: Status 404 returned error can't find the container with id 0b0e03de635e73a91cf5f667c42079d951f513a5f6184502cf203a5c683199c8 Dec 03 00:07:16 crc kubenswrapper[4953]: I1203 00:07:16.589293 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce8b414e1fc22eb30decb6a6411b672e7f443055cb89ebb029fbf6ffc1d5dca7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:16Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:16 crc kubenswrapper[4953]: I1203 00:07:16.602350 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-6dfbn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53ffb1e9-9a2e-4869-80d7-776f4aa7ef8b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:16Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:16Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ct26c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ct26c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:07:16Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-6dfbn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:16Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:16 crc kubenswrapper[4953]: I1203 00:07:16.631299 4953 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-txffx"] Dec 03 00:07:16 crc kubenswrapper[4953]: I1203 00:07:16.632259 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-txffx" Dec 03 00:07:16 crc kubenswrapper[4953]: I1203 00:07:16.640512 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Dec 03 00:07:16 crc kubenswrapper[4953]: I1203 00:07:16.640574 4953 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Dec 03 00:07:16 crc kubenswrapper[4953]: I1203 00:07:16.640626 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Dec 03 00:07:16 crc kubenswrapper[4953]: I1203 00:07:16.640769 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Dec 03 00:07:16 crc kubenswrapper[4953]: I1203 00:07:16.640816 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Dec 03 00:07:16 crc kubenswrapper[4953]: I1203 00:07:16.641027 4953 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Dec 03 00:07:16 crc kubenswrapper[4953]: I1203 00:07:16.641131 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Dec 03 00:07:16 crc kubenswrapper[4953]: I1203 00:07:16.645043 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:16 crc kubenswrapper[4953]: I1203 00:07:16.645074 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:16 crc kubenswrapper[4953]: I1203 00:07:16.645082 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:16 crc kubenswrapper[4953]: I1203 00:07:16.645098 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:16 crc kubenswrapper[4953]: I1203 00:07:16.645107 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:16Z","lastTransitionTime":"2025-12-03T00:07:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:16 crc kubenswrapper[4953]: I1203 00:07:16.655303 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cf821840-5353-4620-8f73-99b56b371ab0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cae5e2994fcc24dc9b2ea73e79572615b3311edc42642b7f278fb84123989b23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://29e4b9d1b70a93fb133e0759ad148ae133766a66efeae3ff06a21dfae0e805ff\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dec43536772aa7c3f5a662fa7f3bd73b1cae2791ad454f85792287828d0ffeed\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://034e0bac6892d361700b8f33d6427b8a55e2bd3f89b14ea77a374c6be94d25e2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://761973393a93c59c23e69f41b8ac0d16abc35abd61ecbc6d51a42d94cd664233\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T00:07:09Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 00:07:03.713374 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 00:07:03.715052 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1528563048/tls.crt::/tmp/serving-cert-1528563048/tls.key\\\\\\\"\\\\nI1203 00:07:09.300052 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 00:07:09.304192 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 00:07:09.304255 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 00:07:09.304309 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 00:07:09.304323 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 00:07:09.313563 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 00:07:09.313597 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 00:07:09.313603 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 00:07:09.313609 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 00:07:09.313613 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 00:07:09.313618 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 00:07:09.313622 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 00:07:09.313709 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 00:07:09.327347 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T00:06:53Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c143d4232fb8bf7ef5999a664e431a00ab0db030f4d31a28d0bb21e00e38581\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:53Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72af6c27296eae2e275d455795c68f0d272869d63642b8353b17d466d0c901cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://72af6c27296eae2e275d455795c68f0d272869d63642b8353b17d466d0c901cf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:06:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:06:51Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:16Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:16 crc kubenswrapper[4953]: I1203 00:07:16.668209 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:10Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:16Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:16 crc kubenswrapper[4953]: I1203 00:07:16.696260 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce8b414e1fc22eb30decb6a6411b672e7f443055cb89ebb029fbf6ffc1d5dca7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:16Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:16 crc kubenswrapper[4953]: I1203 00:07:16.715888 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-6dfbn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53ffb1e9-9a2e-4869-80d7-776f4aa7ef8b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:16Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:16Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ct26c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ct26c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:07:16Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-6dfbn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:16Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:16 crc kubenswrapper[4953]: I1203 00:07:16.744363 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-txffx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8816ebee-c735-4b8d-b1ec-65349707b347\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:16Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:16Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:16Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:07:16Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-txffx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:16Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:16 crc kubenswrapper[4953]: I1203 00:07:16.747794 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:16 crc kubenswrapper[4953]: I1203 00:07:16.747827 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:16 crc kubenswrapper[4953]: I1203 00:07:16.747843 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:16 crc kubenswrapper[4953]: I1203 00:07:16.747870 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:16 crc kubenswrapper[4953]: I1203 00:07:16.747881 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:16Z","lastTransitionTime":"2025-12-03T00:07:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:16 crc kubenswrapper[4953]: I1203 00:07:16.760917 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:16Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:16 crc kubenswrapper[4953]: I1203 00:07:16.761294 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/8816ebee-c735-4b8d-b1ec-65349707b347-host-run-ovn-kubernetes\") pod \"ovnkube-node-txffx\" (UID: \"8816ebee-c735-4b8d-b1ec-65349707b347\") " pod="openshift-ovn-kubernetes/ovnkube-node-txffx" Dec 03 00:07:16 crc kubenswrapper[4953]: I1203 00:07:16.761327 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/8816ebee-c735-4b8d-b1ec-65349707b347-host-cni-netd\") pod \"ovnkube-node-txffx\" (UID: \"8816ebee-c735-4b8d-b1ec-65349707b347\") " pod="openshift-ovn-kubernetes/ovnkube-node-txffx" Dec 03 00:07:16 crc kubenswrapper[4953]: I1203 00:07:16.761354 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/8816ebee-c735-4b8d-b1ec-65349707b347-run-systemd\") pod \"ovnkube-node-txffx\" (UID: \"8816ebee-c735-4b8d-b1ec-65349707b347\") " pod="openshift-ovn-kubernetes/ovnkube-node-txffx" Dec 03 00:07:16 crc kubenswrapper[4953]: I1203 00:07:16.761382 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/8816ebee-c735-4b8d-b1ec-65349707b347-node-log\") pod \"ovnkube-node-txffx\" (UID: \"8816ebee-c735-4b8d-b1ec-65349707b347\") " pod="openshift-ovn-kubernetes/ovnkube-node-txffx" Dec 03 00:07:16 crc kubenswrapper[4953]: I1203 00:07:16.761403 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/8816ebee-c735-4b8d-b1ec-65349707b347-env-overrides\") pod \"ovnkube-node-txffx\" (UID: \"8816ebee-c735-4b8d-b1ec-65349707b347\") " pod="openshift-ovn-kubernetes/ovnkube-node-txffx" Dec 03 00:07:16 crc kubenswrapper[4953]: I1203 00:07:16.761426 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/8816ebee-c735-4b8d-b1ec-65349707b347-systemd-units\") pod \"ovnkube-node-txffx\" (UID: \"8816ebee-c735-4b8d-b1ec-65349707b347\") " pod="openshift-ovn-kubernetes/ovnkube-node-txffx" Dec 03 00:07:16 crc kubenswrapper[4953]: I1203 00:07:16.761445 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/8816ebee-c735-4b8d-b1ec-65349707b347-host-run-netns\") pod \"ovnkube-node-txffx\" (UID: \"8816ebee-c735-4b8d-b1ec-65349707b347\") " pod="openshift-ovn-kubernetes/ovnkube-node-txffx" Dec 03 00:07:16 crc kubenswrapper[4953]: I1203 00:07:16.761463 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/8816ebee-c735-4b8d-b1ec-65349707b347-var-lib-openvswitch\") pod \"ovnkube-node-txffx\" (UID: \"8816ebee-c735-4b8d-b1ec-65349707b347\") " pod="openshift-ovn-kubernetes/ovnkube-node-txffx" Dec 03 00:07:16 crc kubenswrapper[4953]: I1203 00:07:16.761477 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/8816ebee-c735-4b8d-b1ec-65349707b347-log-socket\") pod \"ovnkube-node-txffx\" (UID: \"8816ebee-c735-4b8d-b1ec-65349707b347\") " pod="openshift-ovn-kubernetes/ovnkube-node-txffx" Dec 03 00:07:16 crc kubenswrapper[4953]: I1203 00:07:16.761494 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/8816ebee-c735-4b8d-b1ec-65349707b347-run-ovn\") pod \"ovnkube-node-txffx\" (UID: \"8816ebee-c735-4b8d-b1ec-65349707b347\") " pod="openshift-ovn-kubernetes/ovnkube-node-txffx" Dec 03 00:07:16 crc kubenswrapper[4953]: I1203 00:07:16.761510 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/8816ebee-c735-4b8d-b1ec-65349707b347-ovnkube-script-lib\") pod \"ovnkube-node-txffx\" (UID: \"8816ebee-c735-4b8d-b1ec-65349707b347\") " pod="openshift-ovn-kubernetes/ovnkube-node-txffx" Dec 03 00:07:16 crc kubenswrapper[4953]: I1203 00:07:16.761525 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/8816ebee-c735-4b8d-b1ec-65349707b347-host-cni-bin\") pod \"ovnkube-node-txffx\" (UID: \"8816ebee-c735-4b8d-b1ec-65349707b347\") " pod="openshift-ovn-kubernetes/ovnkube-node-txffx" Dec 03 00:07:16 crc kubenswrapper[4953]: I1203 00:07:16.761542 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/8816ebee-c735-4b8d-b1ec-65349707b347-ovnkube-config\") pod \"ovnkube-node-txffx\" (UID: \"8816ebee-c735-4b8d-b1ec-65349707b347\") " pod="openshift-ovn-kubernetes/ovnkube-node-txffx" Dec 03 00:07:16 crc kubenswrapper[4953]: I1203 00:07:16.761560 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m2fdf\" (UniqueName: \"kubernetes.io/projected/8816ebee-c735-4b8d-b1ec-65349707b347-kube-api-access-m2fdf\") pod \"ovnkube-node-txffx\" (UID: \"8816ebee-c735-4b8d-b1ec-65349707b347\") " pod="openshift-ovn-kubernetes/ovnkube-node-txffx" Dec 03 00:07:16 crc kubenswrapper[4953]: I1203 00:07:16.761577 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/8816ebee-c735-4b8d-b1ec-65349707b347-host-kubelet\") pod \"ovnkube-node-txffx\" (UID: \"8816ebee-c735-4b8d-b1ec-65349707b347\") " pod="openshift-ovn-kubernetes/ovnkube-node-txffx" Dec 03 00:07:16 crc kubenswrapper[4953]: I1203 00:07:16.761597 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/8816ebee-c735-4b8d-b1ec-65349707b347-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-txffx\" (UID: \"8816ebee-c735-4b8d-b1ec-65349707b347\") " pod="openshift-ovn-kubernetes/ovnkube-node-txffx" Dec 03 00:07:16 crc kubenswrapper[4953]: I1203 00:07:16.761630 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/8816ebee-c735-4b8d-b1ec-65349707b347-host-slash\") pod \"ovnkube-node-txffx\" (UID: \"8816ebee-c735-4b8d-b1ec-65349707b347\") " pod="openshift-ovn-kubernetes/ovnkube-node-txffx" Dec 03 00:07:16 crc kubenswrapper[4953]: I1203 00:07:16.761647 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/8816ebee-c735-4b8d-b1ec-65349707b347-etc-openvswitch\") pod \"ovnkube-node-txffx\" (UID: \"8816ebee-c735-4b8d-b1ec-65349707b347\") " pod="openshift-ovn-kubernetes/ovnkube-node-txffx" Dec 03 00:07:16 crc kubenswrapper[4953]: I1203 00:07:16.761664 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/8816ebee-c735-4b8d-b1ec-65349707b347-run-openvswitch\") pod \"ovnkube-node-txffx\" (UID: \"8816ebee-c735-4b8d-b1ec-65349707b347\") " pod="openshift-ovn-kubernetes/ovnkube-node-txffx" Dec 03 00:07:16 crc kubenswrapper[4953]: I1203 00:07:16.761679 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/8816ebee-c735-4b8d-b1ec-65349707b347-ovn-node-metrics-cert\") pod \"ovnkube-node-txffx\" (UID: \"8816ebee-c735-4b8d-b1ec-65349707b347\") " pod="openshift-ovn-kubernetes/ovnkube-node-txffx" Dec 03 00:07:16 crc kubenswrapper[4953]: I1203 00:07:16.777371 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2ed9c7fa96c8951fbc1728fe7a91fefdd999077285621df9c1abedf9ae0c449\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:16Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:16 crc kubenswrapper[4953]: I1203 00:07:16.821066 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:10Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:16Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:16 crc kubenswrapper[4953]: I1203 00:07:16.842815 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-675qk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"43a4bad8-71f4-441a-8302-aeb526a986af\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:16Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:16Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gdrw4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:07:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-675qk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:16Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:16 crc kubenswrapper[4953]: I1203 00:07:16.853168 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:16 crc kubenswrapper[4953]: I1203 00:07:16.853198 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:16 crc kubenswrapper[4953]: I1203 00:07:16.853208 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:16 crc kubenswrapper[4953]: I1203 00:07:16.853224 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:16 crc kubenswrapper[4953]: I1203 00:07:16.853235 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:16Z","lastTransitionTime":"2025-12-03T00:07:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:16 crc kubenswrapper[4953]: I1203 00:07:16.860234 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f94f8c51-c665-4ca3-bf17-7a04ea5c7ade\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72c5cd7dc1c3a8adfedab6ba43a59d80b3efe65f2bb56be95e76cc86a409dcf3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2810ffc1b141fa7da811406053db9d72c708edf5f2bb1dd75e4d52ab35fd24c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c67b4277b20a9f1287f50ef82d0785c9d02543a8e7ed26e0413eae332b866db7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://258d78e1ad9d3dcda6976909ae3fe71aecead5b995fc66b3fd6e2722d80b5ab8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:06:51Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:16Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:16 crc kubenswrapper[4953]: I1203 00:07:16.862361 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/8816ebee-c735-4b8d-b1ec-65349707b347-etc-openvswitch\") pod \"ovnkube-node-txffx\" (UID: \"8816ebee-c735-4b8d-b1ec-65349707b347\") " pod="openshift-ovn-kubernetes/ovnkube-node-txffx" Dec 03 00:07:16 crc kubenswrapper[4953]: I1203 00:07:16.862397 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/8816ebee-c735-4b8d-b1ec-65349707b347-run-openvswitch\") pod \"ovnkube-node-txffx\" (UID: \"8816ebee-c735-4b8d-b1ec-65349707b347\") " pod="openshift-ovn-kubernetes/ovnkube-node-txffx" Dec 03 00:07:16 crc kubenswrapper[4953]: I1203 00:07:16.862416 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/8816ebee-c735-4b8d-b1ec-65349707b347-ovn-node-metrics-cert\") pod \"ovnkube-node-txffx\" (UID: \"8816ebee-c735-4b8d-b1ec-65349707b347\") " pod="openshift-ovn-kubernetes/ovnkube-node-txffx" Dec 03 00:07:16 crc kubenswrapper[4953]: I1203 00:07:16.862436 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/8816ebee-c735-4b8d-b1ec-65349707b347-host-cni-netd\") pod \"ovnkube-node-txffx\" (UID: \"8816ebee-c735-4b8d-b1ec-65349707b347\") " pod="openshift-ovn-kubernetes/ovnkube-node-txffx" Dec 03 00:07:16 crc kubenswrapper[4953]: I1203 00:07:16.862455 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/8816ebee-c735-4b8d-b1ec-65349707b347-host-run-ovn-kubernetes\") pod \"ovnkube-node-txffx\" (UID: \"8816ebee-c735-4b8d-b1ec-65349707b347\") " pod="openshift-ovn-kubernetes/ovnkube-node-txffx" Dec 03 00:07:16 crc kubenswrapper[4953]: I1203 00:07:16.862473 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/8816ebee-c735-4b8d-b1ec-65349707b347-run-systemd\") pod \"ovnkube-node-txffx\" (UID: \"8816ebee-c735-4b8d-b1ec-65349707b347\") " pod="openshift-ovn-kubernetes/ovnkube-node-txffx" Dec 03 00:07:16 crc kubenswrapper[4953]: I1203 00:07:16.862489 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/8816ebee-c735-4b8d-b1ec-65349707b347-node-log\") pod \"ovnkube-node-txffx\" (UID: \"8816ebee-c735-4b8d-b1ec-65349707b347\") " pod="openshift-ovn-kubernetes/ovnkube-node-txffx" Dec 03 00:07:16 crc kubenswrapper[4953]: I1203 00:07:16.862517 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/8816ebee-c735-4b8d-b1ec-65349707b347-env-overrides\") pod \"ovnkube-node-txffx\" (UID: \"8816ebee-c735-4b8d-b1ec-65349707b347\") " pod="openshift-ovn-kubernetes/ovnkube-node-txffx" Dec 03 00:07:16 crc kubenswrapper[4953]: I1203 00:07:16.862519 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/8816ebee-c735-4b8d-b1ec-65349707b347-host-cni-netd\") pod \"ovnkube-node-txffx\" (UID: \"8816ebee-c735-4b8d-b1ec-65349707b347\") " pod="openshift-ovn-kubernetes/ovnkube-node-txffx" Dec 03 00:07:16 crc kubenswrapper[4953]: I1203 00:07:16.862536 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/8816ebee-c735-4b8d-b1ec-65349707b347-systemd-units\") pod \"ovnkube-node-txffx\" (UID: \"8816ebee-c735-4b8d-b1ec-65349707b347\") " pod="openshift-ovn-kubernetes/ovnkube-node-txffx" Dec 03 00:07:16 crc kubenswrapper[4953]: I1203 00:07:16.862574 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/8816ebee-c735-4b8d-b1ec-65349707b347-systemd-units\") pod \"ovnkube-node-txffx\" (UID: \"8816ebee-c735-4b8d-b1ec-65349707b347\") " pod="openshift-ovn-kubernetes/ovnkube-node-txffx" Dec 03 00:07:16 crc kubenswrapper[4953]: I1203 00:07:16.862586 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/8816ebee-c735-4b8d-b1ec-65349707b347-host-run-netns\") pod \"ovnkube-node-txffx\" (UID: \"8816ebee-c735-4b8d-b1ec-65349707b347\") " pod="openshift-ovn-kubernetes/ovnkube-node-txffx" Dec 03 00:07:16 crc kubenswrapper[4953]: I1203 00:07:16.862605 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/8816ebee-c735-4b8d-b1ec-65349707b347-host-run-netns\") pod \"ovnkube-node-txffx\" (UID: \"8816ebee-c735-4b8d-b1ec-65349707b347\") " pod="openshift-ovn-kubernetes/ovnkube-node-txffx" Dec 03 00:07:16 crc kubenswrapper[4953]: I1203 00:07:16.862613 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/8816ebee-c735-4b8d-b1ec-65349707b347-var-lib-openvswitch\") pod \"ovnkube-node-txffx\" (UID: \"8816ebee-c735-4b8d-b1ec-65349707b347\") " pod="openshift-ovn-kubernetes/ovnkube-node-txffx" Dec 03 00:07:16 crc kubenswrapper[4953]: I1203 00:07:16.862628 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/8816ebee-c735-4b8d-b1ec-65349707b347-run-openvswitch\") pod \"ovnkube-node-txffx\" (UID: \"8816ebee-c735-4b8d-b1ec-65349707b347\") " pod="openshift-ovn-kubernetes/ovnkube-node-txffx" Dec 03 00:07:16 crc kubenswrapper[4953]: I1203 00:07:16.862641 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/8816ebee-c735-4b8d-b1ec-65349707b347-log-socket\") pod \"ovnkube-node-txffx\" (UID: \"8816ebee-c735-4b8d-b1ec-65349707b347\") " pod="openshift-ovn-kubernetes/ovnkube-node-txffx" Dec 03 00:07:16 crc kubenswrapper[4953]: I1203 00:07:16.862655 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/8816ebee-c735-4b8d-b1ec-65349707b347-ovnkube-script-lib\") pod \"ovnkube-node-txffx\" (UID: \"8816ebee-c735-4b8d-b1ec-65349707b347\") " pod="openshift-ovn-kubernetes/ovnkube-node-txffx" Dec 03 00:07:16 crc kubenswrapper[4953]: I1203 00:07:16.862670 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/8816ebee-c735-4b8d-b1ec-65349707b347-run-ovn\") pod \"ovnkube-node-txffx\" (UID: \"8816ebee-c735-4b8d-b1ec-65349707b347\") " pod="openshift-ovn-kubernetes/ovnkube-node-txffx" Dec 03 00:07:16 crc kubenswrapper[4953]: I1203 00:07:16.862684 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/8816ebee-c735-4b8d-b1ec-65349707b347-ovnkube-config\") pod \"ovnkube-node-txffx\" (UID: \"8816ebee-c735-4b8d-b1ec-65349707b347\") " pod="openshift-ovn-kubernetes/ovnkube-node-txffx" Dec 03 00:07:16 crc kubenswrapper[4953]: I1203 00:07:16.862701 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m2fdf\" (UniqueName: \"kubernetes.io/projected/8816ebee-c735-4b8d-b1ec-65349707b347-kube-api-access-m2fdf\") pod \"ovnkube-node-txffx\" (UID: \"8816ebee-c735-4b8d-b1ec-65349707b347\") " pod="openshift-ovn-kubernetes/ovnkube-node-txffx" Dec 03 00:07:16 crc kubenswrapper[4953]: I1203 00:07:16.862718 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/8816ebee-c735-4b8d-b1ec-65349707b347-host-cni-bin\") pod \"ovnkube-node-txffx\" (UID: \"8816ebee-c735-4b8d-b1ec-65349707b347\") " pod="openshift-ovn-kubernetes/ovnkube-node-txffx" Dec 03 00:07:16 crc kubenswrapper[4953]: I1203 00:07:16.862736 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/8816ebee-c735-4b8d-b1ec-65349707b347-host-kubelet\") pod \"ovnkube-node-txffx\" (UID: \"8816ebee-c735-4b8d-b1ec-65349707b347\") " pod="openshift-ovn-kubernetes/ovnkube-node-txffx" Dec 03 00:07:16 crc kubenswrapper[4953]: I1203 00:07:16.862753 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/8816ebee-c735-4b8d-b1ec-65349707b347-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-txffx\" (UID: \"8816ebee-c735-4b8d-b1ec-65349707b347\") " pod="openshift-ovn-kubernetes/ovnkube-node-txffx" Dec 03 00:07:16 crc kubenswrapper[4953]: I1203 00:07:16.862769 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/8816ebee-c735-4b8d-b1ec-65349707b347-host-slash\") pod \"ovnkube-node-txffx\" (UID: \"8816ebee-c735-4b8d-b1ec-65349707b347\") " pod="openshift-ovn-kubernetes/ovnkube-node-txffx" Dec 03 00:07:16 crc kubenswrapper[4953]: I1203 00:07:16.862814 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/8816ebee-c735-4b8d-b1ec-65349707b347-host-slash\") pod \"ovnkube-node-txffx\" (UID: \"8816ebee-c735-4b8d-b1ec-65349707b347\") " pod="openshift-ovn-kubernetes/ovnkube-node-txffx" Dec 03 00:07:16 crc kubenswrapper[4953]: I1203 00:07:16.862836 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/8816ebee-c735-4b8d-b1ec-65349707b347-host-run-ovn-kubernetes\") pod \"ovnkube-node-txffx\" (UID: \"8816ebee-c735-4b8d-b1ec-65349707b347\") " pod="openshift-ovn-kubernetes/ovnkube-node-txffx" Dec 03 00:07:16 crc kubenswrapper[4953]: I1203 00:07:16.862856 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/8816ebee-c735-4b8d-b1ec-65349707b347-run-systemd\") pod \"ovnkube-node-txffx\" (UID: \"8816ebee-c735-4b8d-b1ec-65349707b347\") " pod="openshift-ovn-kubernetes/ovnkube-node-txffx" Dec 03 00:07:16 crc kubenswrapper[4953]: I1203 00:07:16.862875 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/8816ebee-c735-4b8d-b1ec-65349707b347-node-log\") pod \"ovnkube-node-txffx\" (UID: \"8816ebee-c735-4b8d-b1ec-65349707b347\") " pod="openshift-ovn-kubernetes/ovnkube-node-txffx" Dec 03 00:07:16 crc kubenswrapper[4953]: I1203 00:07:16.862486 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/8816ebee-c735-4b8d-b1ec-65349707b347-etc-openvswitch\") pod \"ovnkube-node-txffx\" (UID: \"8816ebee-c735-4b8d-b1ec-65349707b347\") " pod="openshift-ovn-kubernetes/ovnkube-node-txffx" Dec 03 00:07:16 crc kubenswrapper[4953]: I1203 00:07:16.863231 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/8816ebee-c735-4b8d-b1ec-65349707b347-host-cni-bin\") pod \"ovnkube-node-txffx\" (UID: \"8816ebee-c735-4b8d-b1ec-65349707b347\") " pod="openshift-ovn-kubernetes/ovnkube-node-txffx" Dec 03 00:07:16 crc kubenswrapper[4953]: I1203 00:07:16.863245 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/8816ebee-c735-4b8d-b1ec-65349707b347-log-socket\") pod \"ovnkube-node-txffx\" (UID: \"8816ebee-c735-4b8d-b1ec-65349707b347\") " pod="openshift-ovn-kubernetes/ovnkube-node-txffx" Dec 03 00:07:16 crc kubenswrapper[4953]: I1203 00:07:16.863294 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/8816ebee-c735-4b8d-b1ec-65349707b347-var-lib-openvswitch\") pod \"ovnkube-node-txffx\" (UID: \"8816ebee-c735-4b8d-b1ec-65349707b347\") " pod="openshift-ovn-kubernetes/ovnkube-node-txffx" Dec 03 00:07:16 crc kubenswrapper[4953]: I1203 00:07:16.863330 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/8816ebee-c735-4b8d-b1ec-65349707b347-host-kubelet\") pod \"ovnkube-node-txffx\" (UID: \"8816ebee-c735-4b8d-b1ec-65349707b347\") " pod="openshift-ovn-kubernetes/ovnkube-node-txffx" Dec 03 00:07:16 crc kubenswrapper[4953]: I1203 00:07:16.863362 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/8816ebee-c735-4b8d-b1ec-65349707b347-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-txffx\" (UID: \"8816ebee-c735-4b8d-b1ec-65349707b347\") " pod="openshift-ovn-kubernetes/ovnkube-node-txffx" Dec 03 00:07:16 crc kubenswrapper[4953]: I1203 00:07:16.863407 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/8816ebee-c735-4b8d-b1ec-65349707b347-env-overrides\") pod \"ovnkube-node-txffx\" (UID: \"8816ebee-c735-4b8d-b1ec-65349707b347\") " pod="openshift-ovn-kubernetes/ovnkube-node-txffx" Dec 03 00:07:16 crc kubenswrapper[4953]: I1203 00:07:16.863440 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/8816ebee-c735-4b8d-b1ec-65349707b347-run-ovn\") pod \"ovnkube-node-txffx\" (UID: \"8816ebee-c735-4b8d-b1ec-65349707b347\") " pod="openshift-ovn-kubernetes/ovnkube-node-txffx" Dec 03 00:07:16 crc kubenswrapper[4953]: I1203 00:07:16.863924 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/8816ebee-c735-4b8d-b1ec-65349707b347-ovnkube-script-lib\") pod \"ovnkube-node-txffx\" (UID: \"8816ebee-c735-4b8d-b1ec-65349707b347\") " pod="openshift-ovn-kubernetes/ovnkube-node-txffx" Dec 03 00:07:16 crc kubenswrapper[4953]: I1203 00:07:16.864355 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/8816ebee-c735-4b8d-b1ec-65349707b347-ovnkube-config\") pod \"ovnkube-node-txffx\" (UID: \"8816ebee-c735-4b8d-b1ec-65349707b347\") " pod="openshift-ovn-kubernetes/ovnkube-node-txffx" Dec 03 00:07:16 crc kubenswrapper[4953]: I1203 00:07:16.867639 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/8816ebee-c735-4b8d-b1ec-65349707b347-ovn-node-metrics-cert\") pod \"ovnkube-node-txffx\" (UID: \"8816ebee-c735-4b8d-b1ec-65349707b347\") " pod="openshift-ovn-kubernetes/ovnkube-node-txffx" Dec 03 00:07:16 crc kubenswrapper[4953]: I1203 00:07:16.882465 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m2fdf\" (UniqueName: \"kubernetes.io/projected/8816ebee-c735-4b8d-b1ec-65349707b347-kube-api-access-m2fdf\") pod \"ovnkube-node-txffx\" (UID: \"8816ebee-c735-4b8d-b1ec-65349707b347\") " pod="openshift-ovn-kubernetes/ovnkube-node-txffx" Dec 03 00:07:16 crc kubenswrapper[4953]: I1203 00:07:16.887349 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09657a9f448a4c92a759c73a0a892076d7eb46393e6fd565e5f73f34c71aac23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5af5eac4a58870f63497f99b503cfa5fb922278095e0332ec50376925ecf2109\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:16Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:16 crc kubenswrapper[4953]: I1203 00:07:16.921638 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-wj6rz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"96aa67a0-15dd-420d-a1d5-ed239d7df1fb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:16Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:16Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:16Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kf62m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kf62m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kf62m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kf62m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kf62m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kf62m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kf62m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:07:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-wj6rz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:16Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:16 crc kubenswrapper[4953]: I1203 00:07:16.939668 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-kl5s9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d85e5c97-16ad-4bd7-9e73-b7f12678be97\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:16Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:16Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j24lb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:07:16Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-kl5s9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:16Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:16 crc kubenswrapper[4953]: I1203 00:07:16.955138 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:16 crc kubenswrapper[4953]: I1203 00:07:16.955186 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:16 crc kubenswrapper[4953]: I1203 00:07:16.955195 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:16 crc kubenswrapper[4953]: I1203 00:07:16.955208 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:16 crc kubenswrapper[4953]: I1203 00:07:16.955221 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:16Z","lastTransitionTime":"2025-12-03T00:07:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:16 crc kubenswrapper[4953]: I1203 00:07:16.961456 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-txffx" Dec 03 00:07:16 crc kubenswrapper[4953]: W1203 00:07:16.973242 4953 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8816ebee_c735_4b8d_b1ec_65349707b347.slice/crio-3780938f7eb5daa186db861efd64edfb729e40aaabab9ded50425004f8f16d42 WatchSource:0}: Error finding container 3780938f7eb5daa186db861efd64edfb729e40aaabab9ded50425004f8f16d42: Status 404 returned error can't find the container with id 3780938f7eb5daa186db861efd64edfb729e40aaabab9ded50425004f8f16d42 Dec 03 00:07:17 crc kubenswrapper[4953]: I1203 00:07:17.057722 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:17 crc kubenswrapper[4953]: I1203 00:07:17.057763 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:17 crc kubenswrapper[4953]: I1203 00:07:17.057779 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:17 crc kubenswrapper[4953]: I1203 00:07:17.057798 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:17 crc kubenswrapper[4953]: I1203 00:07:17.057809 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:17Z","lastTransitionTime":"2025-12-03T00:07:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:17 crc kubenswrapper[4953]: I1203 00:07:17.159681 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:17 crc kubenswrapper[4953]: I1203 00:07:17.159730 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:17 crc kubenswrapper[4953]: I1203 00:07:17.159743 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:17 crc kubenswrapper[4953]: I1203 00:07:17.159758 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:17 crc kubenswrapper[4953]: I1203 00:07:17.159769 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:17Z","lastTransitionTime":"2025-12-03T00:07:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:17 crc kubenswrapper[4953]: I1203 00:07:17.262630 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:17 crc kubenswrapper[4953]: I1203 00:07:17.262691 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:17 crc kubenswrapper[4953]: I1203 00:07:17.262703 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:17 crc kubenswrapper[4953]: I1203 00:07:17.262726 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:17 crc kubenswrapper[4953]: I1203 00:07:17.262743 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:17Z","lastTransitionTime":"2025-12-03T00:07:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:17 crc kubenswrapper[4953]: I1203 00:07:17.368675 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:17 crc kubenswrapper[4953]: I1203 00:07:17.368710 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:17 crc kubenswrapper[4953]: I1203 00:07:17.368719 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:17 crc kubenswrapper[4953]: I1203 00:07:17.368731 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:17 crc kubenswrapper[4953]: I1203 00:07:17.368742 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:17Z","lastTransitionTime":"2025-12-03T00:07:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:17 crc kubenswrapper[4953]: I1203 00:07:17.471757 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:17 crc kubenswrapper[4953]: I1203 00:07:17.471792 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:17 crc kubenswrapper[4953]: I1203 00:07:17.471804 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:17 crc kubenswrapper[4953]: I1203 00:07:17.471824 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:17 crc kubenswrapper[4953]: I1203 00:07:17.471836 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:17Z","lastTransitionTime":"2025-12-03T00:07:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:17 crc kubenswrapper[4953]: I1203 00:07:17.497508 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-6dfbn" event={"ID":"53ffb1e9-9a2e-4869-80d7-776f4aa7ef8b","Type":"ContainerStarted","Data":"ab4205c0d38ad6a0c32077eadb1fd5fd3a279d9c8b8a2ffbe693ec2c4018370c"} Dec 03 00:07:17 crc kubenswrapper[4953]: I1203 00:07:17.497558 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-6dfbn" event={"ID":"53ffb1e9-9a2e-4869-80d7-776f4aa7ef8b","Type":"ContainerStarted","Data":"7df84405eda520ad5a22d35750edf07ef3c3977a756ad5739c54f91cba9da7b2"} Dec 03 00:07:17 crc kubenswrapper[4953]: I1203 00:07:17.497570 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-6dfbn" event={"ID":"53ffb1e9-9a2e-4869-80d7-776f4aa7ef8b","Type":"ContainerStarted","Data":"0b0e03de635e73a91cf5f667c42079d951f513a5f6184502cf203a5c683199c8"} Dec 03 00:07:17 crc kubenswrapper[4953]: I1203 00:07:17.499827 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-675qk" event={"ID":"43a4bad8-71f4-441a-8302-aeb526a986af","Type":"ContainerStarted","Data":"5b5794fef6477aab5deed8f642e8cc50b37eb1fd0cc6f19c003f34ad31c1e3c0"} Dec 03 00:07:17 crc kubenswrapper[4953]: I1203 00:07:17.499855 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-675qk" event={"ID":"43a4bad8-71f4-441a-8302-aeb526a986af","Type":"ContainerStarted","Data":"7de0dff7d485526cf2e1a3c09a3fe9326b95f15c567b0fb543a4cd54ddbaca95"} Dec 03 00:07:17 crc kubenswrapper[4953]: I1203 00:07:17.502373 4953 generic.go:334] "Generic (PLEG): container finished" podID="8816ebee-c735-4b8d-b1ec-65349707b347" containerID="af4c7764fca87144d01f3738ea010308744ccf50c54b7c471ccc213112de00d5" exitCode=0 Dec 03 00:07:17 crc kubenswrapper[4953]: I1203 00:07:17.502482 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-txffx" event={"ID":"8816ebee-c735-4b8d-b1ec-65349707b347","Type":"ContainerDied","Data":"af4c7764fca87144d01f3738ea010308744ccf50c54b7c471ccc213112de00d5"} Dec 03 00:07:17 crc kubenswrapper[4953]: I1203 00:07:17.502577 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-txffx" event={"ID":"8816ebee-c735-4b8d-b1ec-65349707b347","Type":"ContainerStarted","Data":"3780938f7eb5daa186db861efd64edfb729e40aaabab9ded50425004f8f16d42"} Dec 03 00:07:17 crc kubenswrapper[4953]: I1203 00:07:17.503809 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-kl5s9" event={"ID":"d85e5c97-16ad-4bd7-9e73-b7f12678be97","Type":"ContainerStarted","Data":"1c21df91a9e4e3165686882b2eb559498224063dc633c3a63f958598108eca29"} Dec 03 00:07:17 crc kubenswrapper[4953]: I1203 00:07:17.503857 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-kl5s9" event={"ID":"d85e5c97-16ad-4bd7-9e73-b7f12678be97","Type":"ContainerStarted","Data":"f26c42ff35d68d82be39b907391fd7810ac62193c5a72b4b7e5dfab6e8f58bc3"} Dec 03 00:07:17 crc kubenswrapper[4953]: I1203 00:07:17.506176 4953 generic.go:334] "Generic (PLEG): container finished" podID="96aa67a0-15dd-420d-a1d5-ed239d7df1fb" containerID="b12bfb5bdf9fd4f2c50a335fc82df6ee0440eaf8741897da03aa2dae368cc6b4" exitCode=0 Dec 03 00:07:17 crc kubenswrapper[4953]: I1203 00:07:17.506235 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-wj6rz" event={"ID":"96aa67a0-15dd-420d-a1d5-ed239d7df1fb","Type":"ContainerDied","Data":"b12bfb5bdf9fd4f2c50a335fc82df6ee0440eaf8741897da03aa2dae368cc6b4"} Dec 03 00:07:17 crc kubenswrapper[4953]: I1203 00:07:17.506252 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-wj6rz" event={"ID":"96aa67a0-15dd-420d-a1d5-ed239d7df1fb","Type":"ContainerStarted","Data":"1e55d024f1dfd02c91fab473e8d5808cf6546138cedd3a2191b5cdb50782047e"} Dec 03 00:07:17 crc kubenswrapper[4953]: I1203 00:07:17.512407 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cf821840-5353-4620-8f73-99b56b371ab0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cae5e2994fcc24dc9b2ea73e79572615b3311edc42642b7f278fb84123989b23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://29e4b9d1b70a93fb133e0759ad148ae133766a66efeae3ff06a21dfae0e805ff\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dec43536772aa7c3f5a662fa7f3bd73b1cae2791ad454f85792287828d0ffeed\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://034e0bac6892d361700b8f33d6427b8a55e2bd3f89b14ea77a374c6be94d25e2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://761973393a93c59c23e69f41b8ac0d16abc35abd61ecbc6d51a42d94cd664233\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T00:07:09Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 00:07:03.713374 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 00:07:03.715052 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1528563048/tls.crt::/tmp/serving-cert-1528563048/tls.key\\\\\\\"\\\\nI1203 00:07:09.300052 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 00:07:09.304192 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 00:07:09.304255 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 00:07:09.304309 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 00:07:09.304323 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 00:07:09.313563 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 00:07:09.313597 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 00:07:09.313603 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 00:07:09.313609 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 00:07:09.313613 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 00:07:09.313618 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 00:07:09.313622 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 00:07:09.313709 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 00:07:09.327347 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T00:06:53Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c143d4232fb8bf7ef5999a664e431a00ab0db030f4d31a28d0bb21e00e38581\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:53Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72af6c27296eae2e275d455795c68f0d272869d63642b8353b17d466d0c901cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://72af6c27296eae2e275d455795c68f0d272869d63642b8353b17d466d0c901cf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:06:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:06:51Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:17Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:17 crc kubenswrapper[4953]: I1203 00:07:17.538675 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:10Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:17Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:17 crc kubenswrapper[4953]: I1203 00:07:17.551938 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce8b414e1fc22eb30decb6a6411b672e7f443055cb89ebb029fbf6ffc1d5dca7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:17Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:17 crc kubenswrapper[4953]: I1203 00:07:17.565639 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-6dfbn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53ffb1e9-9a2e-4869-80d7-776f4aa7ef8b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ab4205c0d38ad6a0c32077eadb1fd5fd3a279d9c8b8a2ffbe693ec2c4018370c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ct26c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7df84405eda520ad5a22d35750edf07ef3c3977a756ad5739c54f91cba9da7b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ct26c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:07:16Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-6dfbn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:17Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:17 crc kubenswrapper[4953]: I1203 00:07:17.575145 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:17 crc kubenswrapper[4953]: I1203 00:07:17.575177 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:17 crc kubenswrapper[4953]: I1203 00:07:17.575188 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:17 crc kubenswrapper[4953]: I1203 00:07:17.575212 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:17 crc kubenswrapper[4953]: I1203 00:07:17.575223 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:17Z","lastTransitionTime":"2025-12-03T00:07:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:17 crc kubenswrapper[4953]: I1203 00:07:17.584368 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-txffx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8816ebee-c735-4b8d-b1ec-65349707b347\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:16Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:16Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:16Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:07:16Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-txffx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:17Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:17 crc kubenswrapper[4953]: I1203 00:07:17.597378 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:17Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:17 crc kubenswrapper[4953]: I1203 00:07:17.610723 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2ed9c7fa96c8951fbc1728fe7a91fefdd999077285621df9c1abedf9ae0c449\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:17Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:17 crc kubenswrapper[4953]: I1203 00:07:17.624339 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:10Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:17Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:17 crc kubenswrapper[4953]: I1203 00:07:17.637587 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-675qk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"43a4bad8-71f4-441a-8302-aeb526a986af\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:16Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:16Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gdrw4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:07:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-675qk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:17Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:17 crc kubenswrapper[4953]: I1203 00:07:17.651913 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f94f8c51-c665-4ca3-bf17-7a04ea5c7ade\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72c5cd7dc1c3a8adfedab6ba43a59d80b3efe65f2bb56be95e76cc86a409dcf3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2810ffc1b141fa7da811406053db9d72c708edf5f2bb1dd75e4d52ab35fd24c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c67b4277b20a9f1287f50ef82d0785c9d02543a8e7ed26e0413eae332b866db7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://258d78e1ad9d3dcda6976909ae3fe71aecead5b995fc66b3fd6e2722d80b5ab8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:06:51Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:17Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:17 crc kubenswrapper[4953]: I1203 00:07:17.665855 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09657a9f448a4c92a759c73a0a892076d7eb46393e6fd565e5f73f34c71aac23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5af5eac4a58870f63497f99b503cfa5fb922278095e0332ec50376925ecf2109\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:17Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:17 crc kubenswrapper[4953]: I1203 00:07:17.681665 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:17 crc kubenswrapper[4953]: I1203 00:07:17.682000 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:17 crc kubenswrapper[4953]: I1203 00:07:17.682009 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:17 crc kubenswrapper[4953]: I1203 00:07:17.682022 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:17 crc kubenswrapper[4953]: I1203 00:07:17.682030 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:17Z","lastTransitionTime":"2025-12-03T00:07:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:17 crc kubenswrapper[4953]: I1203 00:07:17.684198 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-wj6rz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"96aa67a0-15dd-420d-a1d5-ed239d7df1fb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:16Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:16Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:16Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kf62m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kf62m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kf62m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kf62m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kf62m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kf62m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kf62m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:07:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-wj6rz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:17Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:17 crc kubenswrapper[4953]: I1203 00:07:17.696578 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-kl5s9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d85e5c97-16ad-4bd7-9e73-b7f12678be97\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:16Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:16Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j24lb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:07:16Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-kl5s9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:17Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:17 crc kubenswrapper[4953]: I1203 00:07:17.710594 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:17Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:17 crc kubenswrapper[4953]: I1203 00:07:17.723442 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2ed9c7fa96c8951fbc1728fe7a91fefdd999077285621df9c1abedf9ae0c449\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:17Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:17 crc kubenswrapper[4953]: I1203 00:07:17.736467 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:10Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:17Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:17 crc kubenswrapper[4953]: I1203 00:07:17.748469 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-675qk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"43a4bad8-71f4-441a-8302-aeb526a986af\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5b5794fef6477aab5deed8f642e8cc50b37eb1fd0cc6f19c003f34ad31c1e3c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gdrw4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:07:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-675qk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:17Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:17 crc kubenswrapper[4953]: I1203 00:07:17.758392 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f94f8c51-c665-4ca3-bf17-7a04ea5c7ade\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72c5cd7dc1c3a8adfedab6ba43a59d80b3efe65f2bb56be95e76cc86a409dcf3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2810ffc1b141fa7da811406053db9d72c708edf5f2bb1dd75e4d52ab35fd24c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c67b4277b20a9f1287f50ef82d0785c9d02543a8e7ed26e0413eae332b866db7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://258d78e1ad9d3dcda6976909ae3fe71aecead5b995fc66b3fd6e2722d80b5ab8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:06:51Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:17Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:17 crc kubenswrapper[4953]: I1203 00:07:17.773127 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09657a9f448a4c92a759c73a0a892076d7eb46393e6fd565e5f73f34c71aac23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5af5eac4a58870f63497f99b503cfa5fb922278095e0332ec50376925ecf2109\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:17Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:17 crc kubenswrapper[4953]: I1203 00:07:17.785452 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:17 crc kubenswrapper[4953]: I1203 00:07:17.785495 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:17 crc kubenswrapper[4953]: I1203 00:07:17.785518 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:17 crc kubenswrapper[4953]: I1203 00:07:17.785535 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:17 crc kubenswrapper[4953]: I1203 00:07:17.785546 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:17Z","lastTransitionTime":"2025-12-03T00:07:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:17 crc kubenswrapper[4953]: I1203 00:07:17.789799 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-wj6rz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"96aa67a0-15dd-420d-a1d5-ed239d7df1fb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:16Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:16Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:16Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kf62m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b12bfb5bdf9fd4f2c50a335fc82df6ee0440eaf8741897da03aa2dae368cc6b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b12bfb5bdf9fd4f2c50a335fc82df6ee0440eaf8741897da03aa2dae368cc6b4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:07:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kf62m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kf62m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kf62m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kf62m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kf62m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kf62m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:07:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-wj6rz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:17Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:17 crc kubenswrapper[4953]: I1203 00:07:17.800849 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-kl5s9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d85e5c97-16ad-4bd7-9e73-b7f12678be97\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c21df91a9e4e3165686882b2eb559498224063dc633c3a63f958598108eca29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j24lb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:07:16Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-kl5s9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:17Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:17 crc kubenswrapper[4953]: I1203 00:07:17.813321 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cf821840-5353-4620-8f73-99b56b371ab0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cae5e2994fcc24dc9b2ea73e79572615b3311edc42642b7f278fb84123989b23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://29e4b9d1b70a93fb133e0759ad148ae133766a66efeae3ff06a21dfae0e805ff\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dec43536772aa7c3f5a662fa7f3bd73b1cae2791ad454f85792287828d0ffeed\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://034e0bac6892d361700b8f33d6427b8a55e2bd3f89b14ea77a374c6be94d25e2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://761973393a93c59c23e69f41b8ac0d16abc35abd61ecbc6d51a42d94cd664233\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T00:07:09Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 00:07:03.713374 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 00:07:03.715052 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1528563048/tls.crt::/tmp/serving-cert-1528563048/tls.key\\\\\\\"\\\\nI1203 00:07:09.300052 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 00:07:09.304192 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 00:07:09.304255 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 00:07:09.304309 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 00:07:09.304323 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 00:07:09.313563 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 00:07:09.313597 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 00:07:09.313603 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 00:07:09.313609 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 00:07:09.313613 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 00:07:09.313618 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 00:07:09.313622 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 00:07:09.313709 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 00:07:09.327347 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T00:06:53Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c143d4232fb8bf7ef5999a664e431a00ab0db030f4d31a28d0bb21e00e38581\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:53Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72af6c27296eae2e275d455795c68f0d272869d63642b8353b17d466d0c901cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://72af6c27296eae2e275d455795c68f0d272869d63642b8353b17d466d0c901cf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:06:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:06:51Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:17Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:17 crc kubenswrapper[4953]: I1203 00:07:17.823876 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:10Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:17Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:17 crc kubenswrapper[4953]: I1203 00:07:17.835603 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce8b414e1fc22eb30decb6a6411b672e7f443055cb89ebb029fbf6ffc1d5dca7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:17Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:17 crc kubenswrapper[4953]: I1203 00:07:17.845421 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-6dfbn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53ffb1e9-9a2e-4869-80d7-776f4aa7ef8b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ab4205c0d38ad6a0c32077eadb1fd5fd3a279d9c8b8a2ffbe693ec2c4018370c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ct26c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7df84405eda520ad5a22d35750edf07ef3c3977a756ad5739c54f91cba9da7b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ct26c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:07:16Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-6dfbn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:17Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:17 crc kubenswrapper[4953]: I1203 00:07:17.862457 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-txffx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8816ebee-c735-4b8d-b1ec-65349707b347\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:16Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:16Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af4c7764fca87144d01f3738ea010308744ccf50c54b7c471ccc213112de00d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af4c7764fca87144d01f3738ea010308744ccf50c54b7c471ccc213112de00d5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:07:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:07:16Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-txffx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:17Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:17 crc kubenswrapper[4953]: I1203 00:07:17.888150 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:17 crc kubenswrapper[4953]: I1203 00:07:17.888189 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:17 crc kubenswrapper[4953]: I1203 00:07:17.888200 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:17 crc kubenswrapper[4953]: I1203 00:07:17.888216 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:17 crc kubenswrapper[4953]: I1203 00:07:17.888227 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:17Z","lastTransitionTime":"2025-12-03T00:07:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:17 crc kubenswrapper[4953]: I1203 00:07:17.991021 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:17 crc kubenswrapper[4953]: I1203 00:07:17.991069 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:17 crc kubenswrapper[4953]: I1203 00:07:17.991085 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:17 crc kubenswrapper[4953]: I1203 00:07:17.991101 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:17 crc kubenswrapper[4953]: I1203 00:07:17.991112 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:17Z","lastTransitionTime":"2025-12-03T00:07:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:18 crc kubenswrapper[4953]: I1203 00:07:18.076523 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 00:07:18 crc kubenswrapper[4953]: I1203 00:07:18.076650 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 00:07:18 crc kubenswrapper[4953]: I1203 00:07:18.076701 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 00:07:18 crc kubenswrapper[4953]: I1203 00:07:18.076750 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 00:07:18 crc kubenswrapper[4953]: I1203 00:07:18.076785 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 00:07:18 crc kubenswrapper[4953]: E1203 00:07:18.076934 4953 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 03 00:07:18 crc kubenswrapper[4953]: E1203 00:07:18.076991 4953 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 03 00:07:18 crc kubenswrapper[4953]: E1203 00:07:18.077017 4953 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 03 00:07:18 crc kubenswrapper[4953]: E1203 00:07:18.076936 4953 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 03 00:07:18 crc kubenswrapper[4953]: E1203 00:07:18.077034 4953 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-03 00:07:26.077015793 +0000 UTC m=+35.061296586 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 03 00:07:18 crc kubenswrapper[4953]: E1203 00:07:18.077071 4953 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 03 00:07:18 crc kubenswrapper[4953]: E1203 00:07:18.077074 4953 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 00:07:18 crc kubenswrapper[4953]: E1203 00:07:18.077092 4953 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 00:07:18 crc kubenswrapper[4953]: E1203 00:07:18.077101 4953 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 00:07:26.077070955 +0000 UTC m=+35.061351738 (durationBeforeRetry 8s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 00:07:18 crc kubenswrapper[4953]: E1203 00:07:18.076930 4953 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 03 00:07:18 crc kubenswrapper[4953]: E1203 00:07:18.077142 4953 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-03 00:07:26.077124676 +0000 UTC m=+35.061405549 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 00:07:18 crc kubenswrapper[4953]: E1203 00:07:18.077168 4953 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-03 00:07:26.077157217 +0000 UTC m=+35.061438160 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 00:07:18 crc kubenswrapper[4953]: E1203 00:07:18.077203 4953 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-03 00:07:26.077190368 +0000 UTC m=+35.061471311 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 03 00:07:18 crc kubenswrapper[4953]: I1203 00:07:18.098349 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:18 crc kubenswrapper[4953]: I1203 00:07:18.098637 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:18 crc kubenswrapper[4953]: I1203 00:07:18.098649 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:18 crc kubenswrapper[4953]: I1203 00:07:18.098667 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:18 crc kubenswrapper[4953]: I1203 00:07:18.098679 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:18Z","lastTransitionTime":"2025-12-03T00:07:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:18 crc kubenswrapper[4953]: I1203 00:07:18.189118 4953 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/node-ca-5cmlc"] Dec 03 00:07:18 crc kubenswrapper[4953]: I1203 00:07:18.189587 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-5cmlc" Dec 03 00:07:18 crc kubenswrapper[4953]: I1203 00:07:18.192349 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Dec 03 00:07:18 crc kubenswrapper[4953]: I1203 00:07:18.192504 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Dec 03 00:07:18 crc kubenswrapper[4953]: I1203 00:07:18.192713 4953 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Dec 03 00:07:18 crc kubenswrapper[4953]: I1203 00:07:18.193346 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Dec 03 00:07:18 crc kubenswrapper[4953]: I1203 00:07:18.200926 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:18 crc kubenswrapper[4953]: I1203 00:07:18.201002 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:18 crc kubenswrapper[4953]: I1203 00:07:18.201013 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:18 crc kubenswrapper[4953]: I1203 00:07:18.201027 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:18 crc kubenswrapper[4953]: I1203 00:07:18.201037 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:18Z","lastTransitionTime":"2025-12-03T00:07:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:18 crc kubenswrapper[4953]: I1203 00:07:18.209693 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f94f8c51-c665-4ca3-bf17-7a04ea5c7ade\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72c5cd7dc1c3a8adfedab6ba43a59d80b3efe65f2bb56be95e76cc86a409dcf3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2810ffc1b141fa7da811406053db9d72c708edf5f2bb1dd75e4d52ab35fd24c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c67b4277b20a9f1287f50ef82d0785c9d02543a8e7ed26e0413eae332b866db7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://258d78e1ad9d3dcda6976909ae3fe71aecead5b995fc66b3fd6e2722d80b5ab8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:06:51Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:18Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:18 crc kubenswrapper[4953]: I1203 00:07:18.223520 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09657a9f448a4c92a759c73a0a892076d7eb46393e6fd565e5f73f34c71aac23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5af5eac4a58870f63497f99b503cfa5fb922278095e0332ec50376925ecf2109\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:18Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:18 crc kubenswrapper[4953]: I1203 00:07:18.240751 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-wj6rz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"96aa67a0-15dd-420d-a1d5-ed239d7df1fb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:16Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:16Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:16Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kf62m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b12bfb5bdf9fd4f2c50a335fc82df6ee0440eaf8741897da03aa2dae368cc6b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b12bfb5bdf9fd4f2c50a335fc82df6ee0440eaf8741897da03aa2dae368cc6b4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:07:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kf62m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kf62m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kf62m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kf62m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kf62m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kf62m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:07:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-wj6rz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:18Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:18 crc kubenswrapper[4953]: I1203 00:07:18.253573 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-kl5s9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d85e5c97-16ad-4bd7-9e73-b7f12678be97\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c21df91a9e4e3165686882b2eb559498224063dc633c3a63f958598108eca29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j24lb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:07:16Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-kl5s9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:18Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:18 crc kubenswrapper[4953]: I1203 00:07:18.269742 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cf821840-5353-4620-8f73-99b56b371ab0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cae5e2994fcc24dc9b2ea73e79572615b3311edc42642b7f278fb84123989b23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://29e4b9d1b70a93fb133e0759ad148ae133766a66efeae3ff06a21dfae0e805ff\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dec43536772aa7c3f5a662fa7f3bd73b1cae2791ad454f85792287828d0ffeed\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://034e0bac6892d361700b8f33d6427b8a55e2bd3f89b14ea77a374c6be94d25e2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://761973393a93c59c23e69f41b8ac0d16abc35abd61ecbc6d51a42d94cd664233\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T00:07:09Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 00:07:03.713374 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 00:07:03.715052 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1528563048/tls.crt::/tmp/serving-cert-1528563048/tls.key\\\\\\\"\\\\nI1203 00:07:09.300052 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 00:07:09.304192 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 00:07:09.304255 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 00:07:09.304309 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 00:07:09.304323 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 00:07:09.313563 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 00:07:09.313597 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 00:07:09.313603 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 00:07:09.313609 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 00:07:09.313613 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 00:07:09.313618 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 00:07:09.313622 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 00:07:09.313709 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 00:07:09.327347 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T00:06:53Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c143d4232fb8bf7ef5999a664e431a00ab0db030f4d31a28d0bb21e00e38581\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:53Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72af6c27296eae2e275d455795c68f0d272869d63642b8353b17d466d0c901cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://72af6c27296eae2e275d455795c68f0d272869d63642b8353b17d466d0c901cf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:06:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:06:51Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:18Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:18 crc kubenswrapper[4953]: I1203 00:07:18.278863 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6x2sc\" (UniqueName: \"kubernetes.io/projected/7e11a1d3-ecee-4bf0-87bf-085fab97b73d-kube-api-access-6x2sc\") pod \"node-ca-5cmlc\" (UID: \"7e11a1d3-ecee-4bf0-87bf-085fab97b73d\") " pod="openshift-image-registry/node-ca-5cmlc" Dec 03 00:07:18 crc kubenswrapper[4953]: I1203 00:07:18.278918 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/7e11a1d3-ecee-4bf0-87bf-085fab97b73d-host\") pod \"node-ca-5cmlc\" (UID: \"7e11a1d3-ecee-4bf0-87bf-085fab97b73d\") " pod="openshift-image-registry/node-ca-5cmlc" Dec 03 00:07:18 crc kubenswrapper[4953]: I1203 00:07:18.278959 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/7e11a1d3-ecee-4bf0-87bf-085fab97b73d-serviceca\") pod \"node-ca-5cmlc\" (UID: \"7e11a1d3-ecee-4bf0-87bf-085fab97b73d\") " pod="openshift-image-registry/node-ca-5cmlc" Dec 03 00:07:18 crc kubenswrapper[4953]: I1203 00:07:18.288359 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:10Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:18Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:18 crc kubenswrapper[4953]: I1203 00:07:18.304421 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:18 crc kubenswrapper[4953]: I1203 00:07:18.304480 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:18 crc kubenswrapper[4953]: I1203 00:07:18.304495 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:18 crc kubenswrapper[4953]: I1203 00:07:18.304521 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:18 crc kubenswrapper[4953]: I1203 00:07:18.304536 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:18Z","lastTransitionTime":"2025-12-03T00:07:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:18 crc kubenswrapper[4953]: I1203 00:07:18.308467 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce8b414e1fc22eb30decb6a6411b672e7f443055cb89ebb029fbf6ffc1d5dca7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:18Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:18 crc kubenswrapper[4953]: I1203 00:07:18.331273 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-6dfbn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53ffb1e9-9a2e-4869-80d7-776f4aa7ef8b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ab4205c0d38ad6a0c32077eadb1fd5fd3a279d9c8b8a2ffbe693ec2c4018370c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ct26c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7df84405eda520ad5a22d35750edf07ef3c3977a756ad5739c54f91cba9da7b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ct26c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:07:16Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-6dfbn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:18Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:18 crc kubenswrapper[4953]: I1203 00:07:18.347453 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 00:07:18 crc kubenswrapper[4953]: E1203 00:07:18.347744 4953 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 00:07:18 crc kubenswrapper[4953]: I1203 00:07:18.347791 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 00:07:18 crc kubenswrapper[4953]: I1203 00:07:18.347811 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 00:07:18 crc kubenswrapper[4953]: E1203 00:07:18.347877 4953 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 00:07:18 crc kubenswrapper[4953]: E1203 00:07:18.347968 4953 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 00:07:18 crc kubenswrapper[4953]: I1203 00:07:18.352818 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-txffx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8816ebee-c735-4b8d-b1ec-65349707b347\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:16Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:16Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af4c7764fca87144d01f3738ea010308744ccf50c54b7c471ccc213112de00d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af4c7764fca87144d01f3738ea010308744ccf50c54b7c471ccc213112de00d5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:07:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:07:16Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-txffx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:18Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:18 crc kubenswrapper[4953]: I1203 00:07:18.363605 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5cmlc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7e11a1d3-ecee-4bf0-87bf-085fab97b73d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:18Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:18Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6x2sc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:07:18Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5cmlc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:18Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:18 crc kubenswrapper[4953]: I1203 00:07:18.374724 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:18Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:18 crc kubenswrapper[4953]: I1203 00:07:18.379546 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6x2sc\" (UniqueName: \"kubernetes.io/projected/7e11a1d3-ecee-4bf0-87bf-085fab97b73d-kube-api-access-6x2sc\") pod \"node-ca-5cmlc\" (UID: \"7e11a1d3-ecee-4bf0-87bf-085fab97b73d\") " pod="openshift-image-registry/node-ca-5cmlc" Dec 03 00:07:18 crc kubenswrapper[4953]: I1203 00:07:18.379613 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/7e11a1d3-ecee-4bf0-87bf-085fab97b73d-host\") pod \"node-ca-5cmlc\" (UID: \"7e11a1d3-ecee-4bf0-87bf-085fab97b73d\") " pod="openshift-image-registry/node-ca-5cmlc" Dec 03 00:07:18 crc kubenswrapper[4953]: I1203 00:07:18.379658 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/7e11a1d3-ecee-4bf0-87bf-085fab97b73d-serviceca\") pod \"node-ca-5cmlc\" (UID: \"7e11a1d3-ecee-4bf0-87bf-085fab97b73d\") " pod="openshift-image-registry/node-ca-5cmlc" Dec 03 00:07:18 crc kubenswrapper[4953]: I1203 00:07:18.379810 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/7e11a1d3-ecee-4bf0-87bf-085fab97b73d-host\") pod \"node-ca-5cmlc\" (UID: \"7e11a1d3-ecee-4bf0-87bf-085fab97b73d\") " pod="openshift-image-registry/node-ca-5cmlc" Dec 03 00:07:18 crc kubenswrapper[4953]: I1203 00:07:18.381289 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/7e11a1d3-ecee-4bf0-87bf-085fab97b73d-serviceca\") pod \"node-ca-5cmlc\" (UID: \"7e11a1d3-ecee-4bf0-87bf-085fab97b73d\") " pod="openshift-image-registry/node-ca-5cmlc" Dec 03 00:07:18 crc kubenswrapper[4953]: I1203 00:07:18.387412 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2ed9c7fa96c8951fbc1728fe7a91fefdd999077285621df9c1abedf9ae0c449\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:18Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:18 crc kubenswrapper[4953]: I1203 00:07:18.398572 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:10Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:18Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:18 crc kubenswrapper[4953]: I1203 00:07:18.398765 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6x2sc\" (UniqueName: \"kubernetes.io/projected/7e11a1d3-ecee-4bf0-87bf-085fab97b73d-kube-api-access-6x2sc\") pod \"node-ca-5cmlc\" (UID: \"7e11a1d3-ecee-4bf0-87bf-085fab97b73d\") " pod="openshift-image-registry/node-ca-5cmlc" Dec 03 00:07:18 crc kubenswrapper[4953]: I1203 00:07:18.407839 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:18 crc kubenswrapper[4953]: I1203 00:07:18.407882 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:18 crc kubenswrapper[4953]: I1203 00:07:18.407897 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:18 crc kubenswrapper[4953]: I1203 00:07:18.407918 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:18 crc kubenswrapper[4953]: I1203 00:07:18.407932 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:18Z","lastTransitionTime":"2025-12-03T00:07:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:18 crc kubenswrapper[4953]: I1203 00:07:18.414109 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-675qk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"43a4bad8-71f4-441a-8302-aeb526a986af\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5b5794fef6477aab5deed8f642e8cc50b37eb1fd0cc6f19c003f34ad31c1e3c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gdrw4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:07:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-675qk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:18Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:18 crc kubenswrapper[4953]: I1203 00:07:18.510870 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:18 crc kubenswrapper[4953]: I1203 00:07:18.510913 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:18 crc kubenswrapper[4953]: I1203 00:07:18.510923 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:18 crc kubenswrapper[4953]: I1203 00:07:18.510938 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:18 crc kubenswrapper[4953]: I1203 00:07:18.510956 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:18Z","lastTransitionTime":"2025-12-03T00:07:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:18 crc kubenswrapper[4953]: I1203 00:07:18.513806 4953 generic.go:334] "Generic (PLEG): container finished" podID="96aa67a0-15dd-420d-a1d5-ed239d7df1fb" containerID="83aaa5fa28110bc0ef05aff75cd1442bc9bbdd33b822ae2ef22d4e4ab83238bc" exitCode=0 Dec 03 00:07:18 crc kubenswrapper[4953]: I1203 00:07:18.513893 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-wj6rz" event={"ID":"96aa67a0-15dd-420d-a1d5-ed239d7df1fb","Type":"ContainerDied","Data":"83aaa5fa28110bc0ef05aff75cd1442bc9bbdd33b822ae2ef22d4e4ab83238bc"} Dec 03 00:07:18 crc kubenswrapper[4953]: I1203 00:07:18.521525 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-txffx" event={"ID":"8816ebee-c735-4b8d-b1ec-65349707b347","Type":"ContainerStarted","Data":"e0f6f645a58993f052a6f365bb808f62d99f7fea74886c5cc022b2d86112321a"} Dec 03 00:07:18 crc kubenswrapper[4953]: I1203 00:07:18.521568 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-txffx" event={"ID":"8816ebee-c735-4b8d-b1ec-65349707b347","Type":"ContainerStarted","Data":"d4b85224419a4075acb340e4b11095f2adb5c316e285afa4e19295c2181c0d24"} Dec 03 00:07:18 crc kubenswrapper[4953]: I1203 00:07:18.521581 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-txffx" event={"ID":"8816ebee-c735-4b8d-b1ec-65349707b347","Type":"ContainerStarted","Data":"00a2ea4a9bb870fce30e7fa845f787f184e7a19e42c95511655ea8c5f84a26d1"} Dec 03 00:07:18 crc kubenswrapper[4953]: I1203 00:07:18.521593 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-txffx" event={"ID":"8816ebee-c735-4b8d-b1ec-65349707b347","Type":"ContainerStarted","Data":"dc0a5d4cc9d614d8492626a1b114bdc52f510eb5965d842c02c083f08d4849a7"} Dec 03 00:07:18 crc kubenswrapper[4953]: I1203 00:07:18.531396 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09657a9f448a4c92a759c73a0a892076d7eb46393e6fd565e5f73f34c71aac23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5af5eac4a58870f63497f99b503cfa5fb922278095e0332ec50376925ecf2109\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:18Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:18 crc kubenswrapper[4953]: I1203 00:07:18.540127 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-5cmlc" Dec 03 00:07:18 crc kubenswrapper[4953]: I1203 00:07:18.554327 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-wj6rz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"96aa67a0-15dd-420d-a1d5-ed239d7df1fb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:16Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:16Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:16Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kf62m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b12bfb5bdf9fd4f2c50a335fc82df6ee0440eaf8741897da03aa2dae368cc6b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b12bfb5bdf9fd4f2c50a335fc82df6ee0440eaf8741897da03aa2dae368cc6b4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:07:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kf62m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83aaa5fa28110bc0ef05aff75cd1442bc9bbdd33b822ae2ef22d4e4ab83238bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://83aaa5fa28110bc0ef05aff75cd1442bc9bbdd33b822ae2ef22d4e4ab83238bc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:07:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:07:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kf62m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kf62m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kf62m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kf62m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kf62m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:07:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-wj6rz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:18Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:18 crc kubenswrapper[4953]: I1203 00:07:18.567868 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-kl5s9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d85e5c97-16ad-4bd7-9e73-b7f12678be97\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c21df91a9e4e3165686882b2eb559498224063dc633c3a63f958598108eca29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j24lb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:07:16Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-kl5s9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:18Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:18 crc kubenswrapper[4953]: I1203 00:07:18.587148 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f94f8c51-c665-4ca3-bf17-7a04ea5c7ade\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72c5cd7dc1c3a8adfedab6ba43a59d80b3efe65f2bb56be95e76cc86a409dcf3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2810ffc1b141fa7da811406053db9d72c708edf5f2bb1dd75e4d52ab35fd24c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c67b4277b20a9f1287f50ef82d0785c9d02543a8e7ed26e0413eae332b866db7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://258d78e1ad9d3dcda6976909ae3fe71aecead5b995fc66b3fd6e2722d80b5ab8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:06:51Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:18Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:18 crc kubenswrapper[4953]: I1203 00:07:18.607784 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce8b414e1fc22eb30decb6a6411b672e7f443055cb89ebb029fbf6ffc1d5dca7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:18Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:18 crc kubenswrapper[4953]: I1203 00:07:18.614374 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:18 crc kubenswrapper[4953]: I1203 00:07:18.614411 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:18 crc kubenswrapper[4953]: I1203 00:07:18.614420 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:18 crc kubenswrapper[4953]: I1203 00:07:18.614483 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:18 crc kubenswrapper[4953]: I1203 00:07:18.614495 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:18Z","lastTransitionTime":"2025-12-03T00:07:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:18 crc kubenswrapper[4953]: I1203 00:07:18.621028 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-6dfbn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53ffb1e9-9a2e-4869-80d7-776f4aa7ef8b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ab4205c0d38ad6a0c32077eadb1fd5fd3a279d9c8b8a2ffbe693ec2c4018370c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ct26c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7df84405eda520ad5a22d35750edf07ef3c3977a756ad5739c54f91cba9da7b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ct26c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:07:16Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-6dfbn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:18Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:18 crc kubenswrapper[4953]: I1203 00:07:18.639910 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-txffx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8816ebee-c735-4b8d-b1ec-65349707b347\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:16Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:16Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af4c7764fca87144d01f3738ea010308744ccf50c54b7c471ccc213112de00d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af4c7764fca87144d01f3738ea010308744ccf50c54b7c471ccc213112de00d5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:07:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:07:16Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-txffx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:18Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:18 crc kubenswrapper[4953]: I1203 00:07:18.649807 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5cmlc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7e11a1d3-ecee-4bf0-87bf-085fab97b73d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:18Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:18Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6x2sc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:07:18Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5cmlc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:18Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:18 crc kubenswrapper[4953]: I1203 00:07:18.663512 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cf821840-5353-4620-8f73-99b56b371ab0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cae5e2994fcc24dc9b2ea73e79572615b3311edc42642b7f278fb84123989b23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://29e4b9d1b70a93fb133e0759ad148ae133766a66efeae3ff06a21dfae0e805ff\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dec43536772aa7c3f5a662fa7f3bd73b1cae2791ad454f85792287828d0ffeed\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://034e0bac6892d361700b8f33d6427b8a55e2bd3f89b14ea77a374c6be94d25e2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://761973393a93c59c23e69f41b8ac0d16abc35abd61ecbc6d51a42d94cd664233\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T00:07:09Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 00:07:03.713374 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 00:07:03.715052 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1528563048/tls.crt::/tmp/serving-cert-1528563048/tls.key\\\\\\\"\\\\nI1203 00:07:09.300052 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 00:07:09.304192 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 00:07:09.304255 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 00:07:09.304309 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 00:07:09.304323 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 00:07:09.313563 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 00:07:09.313597 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 00:07:09.313603 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 00:07:09.313609 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 00:07:09.313613 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 00:07:09.313618 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 00:07:09.313622 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 00:07:09.313709 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 00:07:09.327347 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T00:06:53Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c143d4232fb8bf7ef5999a664e431a00ab0db030f4d31a28d0bb21e00e38581\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:53Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72af6c27296eae2e275d455795c68f0d272869d63642b8353b17d466d0c901cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://72af6c27296eae2e275d455795c68f0d272869d63642b8353b17d466d0c901cf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:06:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:06:51Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:18Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:18 crc kubenswrapper[4953]: I1203 00:07:18.682429 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:10Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:18Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:18 crc kubenswrapper[4953]: I1203 00:07:18.696962 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:18Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:18 crc kubenswrapper[4953]: I1203 00:07:18.715160 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2ed9c7fa96c8951fbc1728fe7a91fefdd999077285621df9c1abedf9ae0c449\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:18Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:18 crc kubenswrapper[4953]: I1203 00:07:18.716893 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:18 crc kubenswrapper[4953]: I1203 00:07:18.716940 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:18 crc kubenswrapper[4953]: I1203 00:07:18.716953 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:18 crc kubenswrapper[4953]: I1203 00:07:18.716993 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:18 crc kubenswrapper[4953]: I1203 00:07:18.717010 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:18Z","lastTransitionTime":"2025-12-03T00:07:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:18 crc kubenswrapper[4953]: I1203 00:07:18.727616 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:10Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:18Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:18 crc kubenswrapper[4953]: I1203 00:07:18.751415 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-675qk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"43a4bad8-71f4-441a-8302-aeb526a986af\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5b5794fef6477aab5deed8f642e8cc50b37eb1fd0cc6f19c003f34ad31c1e3c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gdrw4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:07:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-675qk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:18Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:18 crc kubenswrapper[4953]: I1203 00:07:18.819711 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:18 crc kubenswrapper[4953]: I1203 00:07:18.819753 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:18 crc kubenswrapper[4953]: I1203 00:07:18.819764 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:18 crc kubenswrapper[4953]: I1203 00:07:18.819783 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:18 crc kubenswrapper[4953]: I1203 00:07:18.819796 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:18Z","lastTransitionTime":"2025-12-03T00:07:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:18 crc kubenswrapper[4953]: I1203 00:07:18.922757 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:18 crc kubenswrapper[4953]: I1203 00:07:18.922813 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:18 crc kubenswrapper[4953]: I1203 00:07:18.922824 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:18 crc kubenswrapper[4953]: I1203 00:07:18.922846 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:18 crc kubenswrapper[4953]: I1203 00:07:18.922857 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:18Z","lastTransitionTime":"2025-12-03T00:07:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:19 crc kubenswrapper[4953]: I1203 00:07:19.026369 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:19 crc kubenswrapper[4953]: I1203 00:07:19.026953 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:19 crc kubenswrapper[4953]: I1203 00:07:19.027023 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:19 crc kubenswrapper[4953]: I1203 00:07:19.027058 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:19 crc kubenswrapper[4953]: I1203 00:07:19.027087 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:19Z","lastTransitionTime":"2025-12-03T00:07:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:19 crc kubenswrapper[4953]: I1203 00:07:19.130589 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:19 crc kubenswrapper[4953]: I1203 00:07:19.130646 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:19 crc kubenswrapper[4953]: I1203 00:07:19.130701 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:19 crc kubenswrapper[4953]: I1203 00:07:19.130728 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:19 crc kubenswrapper[4953]: I1203 00:07:19.130746 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:19Z","lastTransitionTime":"2025-12-03T00:07:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:19 crc kubenswrapper[4953]: I1203 00:07:19.233690 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:19 crc kubenswrapper[4953]: I1203 00:07:19.233774 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:19 crc kubenswrapper[4953]: I1203 00:07:19.233798 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:19 crc kubenswrapper[4953]: I1203 00:07:19.233833 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:19 crc kubenswrapper[4953]: I1203 00:07:19.233860 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:19Z","lastTransitionTime":"2025-12-03T00:07:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:19 crc kubenswrapper[4953]: I1203 00:07:19.337595 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:19 crc kubenswrapper[4953]: I1203 00:07:19.337667 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:19 crc kubenswrapper[4953]: I1203 00:07:19.337687 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:19 crc kubenswrapper[4953]: I1203 00:07:19.337715 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:19 crc kubenswrapper[4953]: I1203 00:07:19.337735 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:19Z","lastTransitionTime":"2025-12-03T00:07:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:19 crc kubenswrapper[4953]: I1203 00:07:19.441185 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:19 crc kubenswrapper[4953]: I1203 00:07:19.441254 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:19 crc kubenswrapper[4953]: I1203 00:07:19.441269 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:19 crc kubenswrapper[4953]: I1203 00:07:19.441297 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:19 crc kubenswrapper[4953]: I1203 00:07:19.441313 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:19Z","lastTransitionTime":"2025-12-03T00:07:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:19 crc kubenswrapper[4953]: I1203 00:07:19.531811 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-5cmlc" event={"ID":"7e11a1d3-ecee-4bf0-87bf-085fab97b73d","Type":"ContainerStarted","Data":"d42390b9c53efc63add8ddcebf03c976f5e5448eaa2c34e076b2dc90dbac60fa"} Dec 03 00:07:19 crc kubenswrapper[4953]: I1203 00:07:19.531873 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-5cmlc" event={"ID":"7e11a1d3-ecee-4bf0-87bf-085fab97b73d","Type":"ContainerStarted","Data":"62cbea432c3ab20b6d5e1d6736bd4392ef38797a0b751acbc5d2a78cc15f0589"} Dec 03 00:07:19 crc kubenswrapper[4953]: I1203 00:07:19.537937 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-txffx" event={"ID":"8816ebee-c735-4b8d-b1ec-65349707b347","Type":"ContainerStarted","Data":"a61b951a9cbab2c48c0f2df8351f97d7ff71bf3674f641fe5eac30155232b224"} Dec 03 00:07:19 crc kubenswrapper[4953]: I1203 00:07:19.538023 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-txffx" event={"ID":"8816ebee-c735-4b8d-b1ec-65349707b347","Type":"ContainerStarted","Data":"fd5f428de229749485267e81d34e31b708728b52474032e88460e54747f21e11"} Dec 03 00:07:19 crc kubenswrapper[4953]: I1203 00:07:19.540877 4953 generic.go:334] "Generic (PLEG): container finished" podID="96aa67a0-15dd-420d-a1d5-ed239d7df1fb" containerID="3ab6213a33a19fedebce6d74a0ebbb3b3a2de5687504e21cc24093cc42ade635" exitCode=0 Dec 03 00:07:19 crc kubenswrapper[4953]: I1203 00:07:19.540931 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-wj6rz" event={"ID":"96aa67a0-15dd-420d-a1d5-ed239d7df1fb","Type":"ContainerDied","Data":"3ab6213a33a19fedebce6d74a0ebbb3b3a2de5687504e21cc24093cc42ade635"} Dec 03 00:07:19 crc kubenswrapper[4953]: I1203 00:07:19.544592 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:19 crc kubenswrapper[4953]: I1203 00:07:19.544656 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:19 crc kubenswrapper[4953]: I1203 00:07:19.544679 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:19 crc kubenswrapper[4953]: I1203 00:07:19.544707 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:19 crc kubenswrapper[4953]: I1203 00:07:19.544733 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:19Z","lastTransitionTime":"2025-12-03T00:07:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:19 crc kubenswrapper[4953]: I1203 00:07:19.553682 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce8b414e1fc22eb30decb6a6411b672e7f443055cb89ebb029fbf6ffc1d5dca7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:19Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:19 crc kubenswrapper[4953]: I1203 00:07:19.570177 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-6dfbn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53ffb1e9-9a2e-4869-80d7-776f4aa7ef8b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ab4205c0d38ad6a0c32077eadb1fd5fd3a279d9c8b8a2ffbe693ec2c4018370c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ct26c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7df84405eda520ad5a22d35750edf07ef3c3977a756ad5739c54f91cba9da7b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ct26c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:07:16Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-6dfbn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:19Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:19 crc kubenswrapper[4953]: I1203 00:07:19.603396 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-txffx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8816ebee-c735-4b8d-b1ec-65349707b347\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:16Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:16Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af4c7764fca87144d01f3738ea010308744ccf50c54b7c471ccc213112de00d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af4c7764fca87144d01f3738ea010308744ccf50c54b7c471ccc213112de00d5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:07:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:07:16Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-txffx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:19Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:19 crc kubenswrapper[4953]: I1203 00:07:19.617319 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5cmlc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7e11a1d3-ecee-4bf0-87bf-085fab97b73d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d42390b9c53efc63add8ddcebf03c976f5e5448eaa2c34e076b2dc90dbac60fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6x2sc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:07:18Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5cmlc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:19Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:19 crc kubenswrapper[4953]: I1203 00:07:19.633659 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cf821840-5353-4620-8f73-99b56b371ab0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cae5e2994fcc24dc9b2ea73e79572615b3311edc42642b7f278fb84123989b23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://29e4b9d1b70a93fb133e0759ad148ae133766a66efeae3ff06a21dfae0e805ff\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dec43536772aa7c3f5a662fa7f3bd73b1cae2791ad454f85792287828d0ffeed\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://034e0bac6892d361700b8f33d6427b8a55e2bd3f89b14ea77a374c6be94d25e2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://761973393a93c59c23e69f41b8ac0d16abc35abd61ecbc6d51a42d94cd664233\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T00:07:09Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 00:07:03.713374 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 00:07:03.715052 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1528563048/tls.crt::/tmp/serving-cert-1528563048/tls.key\\\\\\\"\\\\nI1203 00:07:09.300052 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 00:07:09.304192 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 00:07:09.304255 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 00:07:09.304309 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 00:07:09.304323 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 00:07:09.313563 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 00:07:09.313597 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 00:07:09.313603 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 00:07:09.313609 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 00:07:09.313613 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 00:07:09.313618 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 00:07:09.313622 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 00:07:09.313709 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 00:07:09.327347 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T00:06:53Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c143d4232fb8bf7ef5999a664e431a00ab0db030f4d31a28d0bb21e00e38581\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:53Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72af6c27296eae2e275d455795c68f0d272869d63642b8353b17d466d0c901cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://72af6c27296eae2e275d455795c68f0d272869d63642b8353b17d466d0c901cf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:06:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:06:51Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:19Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:19 crc kubenswrapper[4953]: I1203 00:07:19.651331 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:19 crc kubenswrapper[4953]: I1203 00:07:19.651397 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:19 crc kubenswrapper[4953]: I1203 00:07:19.651416 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:19 crc kubenswrapper[4953]: I1203 00:07:19.651441 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:19 crc kubenswrapper[4953]: I1203 00:07:19.651457 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:19Z","lastTransitionTime":"2025-12-03T00:07:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:19 crc kubenswrapper[4953]: I1203 00:07:19.658417 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:10Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:19Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:19 crc kubenswrapper[4953]: I1203 00:07:19.679422 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:19Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:19 crc kubenswrapper[4953]: I1203 00:07:19.698280 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2ed9c7fa96c8951fbc1728fe7a91fefdd999077285621df9c1abedf9ae0c449\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:19Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:19 crc kubenswrapper[4953]: I1203 00:07:19.717057 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:10Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:19Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:19 crc kubenswrapper[4953]: I1203 00:07:19.745432 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-675qk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"43a4bad8-71f4-441a-8302-aeb526a986af\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5b5794fef6477aab5deed8f642e8cc50b37eb1fd0cc6f19c003f34ad31c1e3c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gdrw4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:07:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-675qk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:19Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:19 crc kubenswrapper[4953]: I1203 00:07:19.759735 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:19 crc kubenswrapper[4953]: I1203 00:07:19.759768 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:19 crc kubenswrapper[4953]: I1203 00:07:19.759779 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:19 crc kubenswrapper[4953]: I1203 00:07:19.759795 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:19 crc kubenswrapper[4953]: I1203 00:07:19.759808 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:19Z","lastTransitionTime":"2025-12-03T00:07:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:19 crc kubenswrapper[4953]: I1203 00:07:19.771096 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09657a9f448a4c92a759c73a0a892076d7eb46393e6fd565e5f73f34c71aac23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5af5eac4a58870f63497f99b503cfa5fb922278095e0332ec50376925ecf2109\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:19Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:19 crc kubenswrapper[4953]: I1203 00:07:19.799022 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-wj6rz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"96aa67a0-15dd-420d-a1d5-ed239d7df1fb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:16Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:16Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:16Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kf62m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b12bfb5bdf9fd4f2c50a335fc82df6ee0440eaf8741897da03aa2dae368cc6b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b12bfb5bdf9fd4f2c50a335fc82df6ee0440eaf8741897da03aa2dae368cc6b4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:07:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kf62m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83aaa5fa28110bc0ef05aff75cd1442bc9bbdd33b822ae2ef22d4e4ab83238bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://83aaa5fa28110bc0ef05aff75cd1442bc9bbdd33b822ae2ef22d4e4ab83238bc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:07:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:07:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kf62m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kf62m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kf62m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kf62m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kf62m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:07:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-wj6rz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:19Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:19 crc kubenswrapper[4953]: I1203 00:07:19.821339 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-kl5s9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d85e5c97-16ad-4bd7-9e73-b7f12678be97\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c21df91a9e4e3165686882b2eb559498224063dc633c3a63f958598108eca29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j24lb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:07:16Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-kl5s9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:19Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:19 crc kubenswrapper[4953]: I1203 00:07:19.836694 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f94f8c51-c665-4ca3-bf17-7a04ea5c7ade\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72c5cd7dc1c3a8adfedab6ba43a59d80b3efe65f2bb56be95e76cc86a409dcf3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2810ffc1b141fa7da811406053db9d72c708edf5f2bb1dd75e4d52ab35fd24c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c67b4277b20a9f1287f50ef82d0785c9d02543a8e7ed26e0413eae332b866db7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://258d78e1ad9d3dcda6976909ae3fe71aecead5b995fc66b3fd6e2722d80b5ab8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:06:51Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:19Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:19 crc kubenswrapper[4953]: I1203 00:07:19.860414 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5cmlc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7e11a1d3-ecee-4bf0-87bf-085fab97b73d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d42390b9c53efc63add8ddcebf03c976f5e5448eaa2c34e076b2dc90dbac60fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6x2sc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:07:18Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5cmlc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:19Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:19 crc kubenswrapper[4953]: I1203 00:07:19.862331 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:19 crc kubenswrapper[4953]: I1203 00:07:19.862372 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:19 crc kubenswrapper[4953]: I1203 00:07:19.862384 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:19 crc kubenswrapper[4953]: I1203 00:07:19.862407 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:19 crc kubenswrapper[4953]: I1203 00:07:19.862421 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:19Z","lastTransitionTime":"2025-12-03T00:07:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:19 crc kubenswrapper[4953]: I1203 00:07:19.899455 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cf821840-5353-4620-8f73-99b56b371ab0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cae5e2994fcc24dc9b2ea73e79572615b3311edc42642b7f278fb84123989b23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://29e4b9d1b70a93fb133e0759ad148ae133766a66efeae3ff06a21dfae0e805ff\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dec43536772aa7c3f5a662fa7f3bd73b1cae2791ad454f85792287828d0ffeed\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://034e0bac6892d361700b8f33d6427b8a55e2bd3f89b14ea77a374c6be94d25e2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://761973393a93c59c23e69f41b8ac0d16abc35abd61ecbc6d51a42d94cd664233\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T00:07:09Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 00:07:03.713374 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 00:07:03.715052 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1528563048/tls.crt::/tmp/serving-cert-1528563048/tls.key\\\\\\\"\\\\nI1203 00:07:09.300052 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 00:07:09.304192 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 00:07:09.304255 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 00:07:09.304309 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 00:07:09.304323 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 00:07:09.313563 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 00:07:09.313597 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 00:07:09.313603 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 00:07:09.313609 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 00:07:09.313613 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 00:07:09.313618 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 00:07:09.313622 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 00:07:09.313709 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 00:07:09.327347 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T00:06:53Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c143d4232fb8bf7ef5999a664e431a00ab0db030f4d31a28d0bb21e00e38581\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:53Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72af6c27296eae2e275d455795c68f0d272869d63642b8353b17d466d0c901cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://72af6c27296eae2e275d455795c68f0d272869d63642b8353b17d466d0c901cf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:06:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:06:51Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:19Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:19 crc kubenswrapper[4953]: I1203 00:07:19.938207 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:10Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:19Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:19 crc kubenswrapper[4953]: I1203 00:07:19.960915 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce8b414e1fc22eb30decb6a6411b672e7f443055cb89ebb029fbf6ffc1d5dca7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:19Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:19 crc kubenswrapper[4953]: I1203 00:07:19.964319 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:19 crc kubenswrapper[4953]: I1203 00:07:19.964354 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:19 crc kubenswrapper[4953]: I1203 00:07:19.964362 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:19 crc kubenswrapper[4953]: I1203 00:07:19.964378 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:19 crc kubenswrapper[4953]: I1203 00:07:19.964387 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:19Z","lastTransitionTime":"2025-12-03T00:07:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:19 crc kubenswrapper[4953]: I1203 00:07:19.977778 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-6dfbn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53ffb1e9-9a2e-4869-80d7-776f4aa7ef8b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ab4205c0d38ad6a0c32077eadb1fd5fd3a279d9c8b8a2ffbe693ec2c4018370c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ct26c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7df84405eda520ad5a22d35750edf07ef3c3977a756ad5739c54f91cba9da7b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ct26c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:07:16Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-6dfbn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:19Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:19 crc kubenswrapper[4953]: I1203 00:07:19.997872 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-txffx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8816ebee-c735-4b8d-b1ec-65349707b347\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:16Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:16Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af4c7764fca87144d01f3738ea010308744ccf50c54b7c471ccc213112de00d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af4c7764fca87144d01f3738ea010308744ccf50c54b7c471ccc213112de00d5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:07:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:07:16Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-txffx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:19Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:20 crc kubenswrapper[4953]: I1203 00:07:20.010588 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:20Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:20 crc kubenswrapper[4953]: I1203 00:07:20.020832 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2ed9c7fa96c8951fbc1728fe7a91fefdd999077285621df9c1abedf9ae0c449\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:20Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:20 crc kubenswrapper[4953]: I1203 00:07:20.033568 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:10Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:20Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:20 crc kubenswrapper[4953]: I1203 00:07:20.048553 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-675qk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"43a4bad8-71f4-441a-8302-aeb526a986af\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5b5794fef6477aab5deed8f642e8cc50b37eb1fd0cc6f19c003f34ad31c1e3c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gdrw4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:07:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-675qk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:20Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:20 crc kubenswrapper[4953]: I1203 00:07:20.064437 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f94f8c51-c665-4ca3-bf17-7a04ea5c7ade\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72c5cd7dc1c3a8adfedab6ba43a59d80b3efe65f2bb56be95e76cc86a409dcf3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2810ffc1b141fa7da811406053db9d72c708edf5f2bb1dd75e4d52ab35fd24c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c67b4277b20a9f1287f50ef82d0785c9d02543a8e7ed26e0413eae332b866db7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://258d78e1ad9d3dcda6976909ae3fe71aecead5b995fc66b3fd6e2722d80b5ab8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:06:51Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:20Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:20 crc kubenswrapper[4953]: I1203 00:07:20.066607 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:20 crc kubenswrapper[4953]: I1203 00:07:20.066657 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:20 crc kubenswrapper[4953]: I1203 00:07:20.066668 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:20 crc kubenswrapper[4953]: I1203 00:07:20.066690 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:20 crc kubenswrapper[4953]: I1203 00:07:20.066704 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:20Z","lastTransitionTime":"2025-12-03T00:07:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:20 crc kubenswrapper[4953]: I1203 00:07:20.082645 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09657a9f448a4c92a759c73a0a892076d7eb46393e6fd565e5f73f34c71aac23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5af5eac4a58870f63497f99b503cfa5fb922278095e0332ec50376925ecf2109\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:20Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:20 crc kubenswrapper[4953]: I1203 00:07:20.103266 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-wj6rz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"96aa67a0-15dd-420d-a1d5-ed239d7df1fb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:16Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:16Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:16Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kf62m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b12bfb5bdf9fd4f2c50a335fc82df6ee0440eaf8741897da03aa2dae368cc6b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b12bfb5bdf9fd4f2c50a335fc82df6ee0440eaf8741897da03aa2dae368cc6b4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:07:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kf62m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83aaa5fa28110bc0ef05aff75cd1442bc9bbdd33b822ae2ef22d4e4ab83238bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://83aaa5fa28110bc0ef05aff75cd1442bc9bbdd33b822ae2ef22d4e4ab83238bc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:07:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:07:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kf62m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ab6213a33a19fedebce6d74a0ebbb3b3a2de5687504e21cc24093cc42ade635\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ab6213a33a19fedebce6d74a0ebbb3b3a2de5687504e21cc24093cc42ade635\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:07:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:07:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kf62m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kf62m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kf62m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kf62m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:07:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-wj6rz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:20Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:20 crc kubenswrapper[4953]: I1203 00:07:20.115269 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-kl5s9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d85e5c97-16ad-4bd7-9e73-b7f12678be97\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c21df91a9e4e3165686882b2eb559498224063dc633c3a63f958598108eca29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j24lb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:07:16Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-kl5s9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:20Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:20 crc kubenswrapper[4953]: I1203 00:07:20.169910 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:20 crc kubenswrapper[4953]: I1203 00:07:20.169993 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:20 crc kubenswrapper[4953]: I1203 00:07:20.170006 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:20 crc kubenswrapper[4953]: I1203 00:07:20.170032 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:20 crc kubenswrapper[4953]: I1203 00:07:20.170045 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:20Z","lastTransitionTime":"2025-12-03T00:07:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:20 crc kubenswrapper[4953]: I1203 00:07:20.273433 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:20 crc kubenswrapper[4953]: I1203 00:07:20.273515 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:20 crc kubenswrapper[4953]: I1203 00:07:20.273529 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:20 crc kubenswrapper[4953]: I1203 00:07:20.273554 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:20 crc kubenswrapper[4953]: I1203 00:07:20.273568 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:20Z","lastTransitionTime":"2025-12-03T00:07:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:20 crc kubenswrapper[4953]: I1203 00:07:20.347232 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 00:07:20 crc kubenswrapper[4953]: I1203 00:07:20.347395 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 00:07:20 crc kubenswrapper[4953]: I1203 00:07:20.347611 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 00:07:20 crc kubenswrapper[4953]: E1203 00:07:20.347613 4953 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 00:07:20 crc kubenswrapper[4953]: E1203 00:07:20.347729 4953 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 00:07:20 crc kubenswrapper[4953]: E1203 00:07:20.347895 4953 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 00:07:20 crc kubenswrapper[4953]: I1203 00:07:20.376958 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:20 crc kubenswrapper[4953]: I1203 00:07:20.377102 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:20 crc kubenswrapper[4953]: I1203 00:07:20.377132 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:20 crc kubenswrapper[4953]: I1203 00:07:20.377167 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:20 crc kubenswrapper[4953]: I1203 00:07:20.377193 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:20Z","lastTransitionTime":"2025-12-03T00:07:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:20 crc kubenswrapper[4953]: I1203 00:07:20.480894 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:20 crc kubenswrapper[4953]: I1203 00:07:20.480949 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:20 crc kubenswrapper[4953]: I1203 00:07:20.480961 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:20 crc kubenswrapper[4953]: I1203 00:07:20.480994 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:20 crc kubenswrapper[4953]: I1203 00:07:20.481008 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:20Z","lastTransitionTime":"2025-12-03T00:07:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:20 crc kubenswrapper[4953]: I1203 00:07:20.549244 4953 generic.go:334] "Generic (PLEG): container finished" podID="96aa67a0-15dd-420d-a1d5-ed239d7df1fb" containerID="de480488843afa6b8dede1e1b42b5181dca1993e2f39dc1b461a18a57e3668e1" exitCode=0 Dec 03 00:07:20 crc kubenswrapper[4953]: I1203 00:07:20.549357 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-wj6rz" event={"ID":"96aa67a0-15dd-420d-a1d5-ed239d7df1fb","Type":"ContainerDied","Data":"de480488843afa6b8dede1e1b42b5181dca1993e2f39dc1b461a18a57e3668e1"} Dec 03 00:07:20 crc kubenswrapper[4953]: I1203 00:07:20.565829 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:20Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:20 crc kubenswrapper[4953]: I1203 00:07:20.583466 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2ed9c7fa96c8951fbc1728fe7a91fefdd999077285621df9c1abedf9ae0c449\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:20Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:20 crc kubenswrapper[4953]: I1203 00:07:20.584564 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:20 crc kubenswrapper[4953]: I1203 00:07:20.584606 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:20 crc kubenswrapper[4953]: I1203 00:07:20.584620 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:20 crc kubenswrapper[4953]: I1203 00:07:20.584640 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:20 crc kubenswrapper[4953]: I1203 00:07:20.584656 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:20Z","lastTransitionTime":"2025-12-03T00:07:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:20 crc kubenswrapper[4953]: I1203 00:07:20.601458 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:10Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:20Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:20 crc kubenswrapper[4953]: I1203 00:07:20.618446 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-675qk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"43a4bad8-71f4-441a-8302-aeb526a986af\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5b5794fef6477aab5deed8f642e8cc50b37eb1fd0cc6f19c003f34ad31c1e3c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gdrw4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:07:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-675qk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:20Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:20 crc kubenswrapper[4953]: I1203 00:07:20.631833 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f94f8c51-c665-4ca3-bf17-7a04ea5c7ade\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72c5cd7dc1c3a8adfedab6ba43a59d80b3efe65f2bb56be95e76cc86a409dcf3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2810ffc1b141fa7da811406053db9d72c708edf5f2bb1dd75e4d52ab35fd24c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c67b4277b20a9f1287f50ef82d0785c9d02543a8e7ed26e0413eae332b866db7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://258d78e1ad9d3dcda6976909ae3fe71aecead5b995fc66b3fd6e2722d80b5ab8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:06:51Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:20Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:20 crc kubenswrapper[4953]: I1203 00:07:20.644358 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09657a9f448a4c92a759c73a0a892076d7eb46393e6fd565e5f73f34c71aac23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5af5eac4a58870f63497f99b503cfa5fb922278095e0332ec50376925ecf2109\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:20Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:20 crc kubenswrapper[4953]: I1203 00:07:20.660328 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-wj6rz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"96aa67a0-15dd-420d-a1d5-ed239d7df1fb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:16Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:16Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:16Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kf62m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b12bfb5bdf9fd4f2c50a335fc82df6ee0440eaf8741897da03aa2dae368cc6b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b12bfb5bdf9fd4f2c50a335fc82df6ee0440eaf8741897da03aa2dae368cc6b4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:07:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kf62m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83aaa5fa28110bc0ef05aff75cd1442bc9bbdd33b822ae2ef22d4e4ab83238bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://83aaa5fa28110bc0ef05aff75cd1442bc9bbdd33b822ae2ef22d4e4ab83238bc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:07:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:07:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kf62m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ab6213a33a19fedebce6d74a0ebbb3b3a2de5687504e21cc24093cc42ade635\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ab6213a33a19fedebce6d74a0ebbb3b3a2de5687504e21cc24093cc42ade635\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:07:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:07:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kf62m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://de480488843afa6b8dede1e1b42b5181dca1993e2f39dc1b461a18a57e3668e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://de480488843afa6b8dede1e1b42b5181dca1993e2f39dc1b461a18a57e3668e1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:07:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kf62m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kf62m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kf62m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:07:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-wj6rz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:20Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:20 crc kubenswrapper[4953]: I1203 00:07:20.671936 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-kl5s9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d85e5c97-16ad-4bd7-9e73-b7f12678be97\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c21df91a9e4e3165686882b2eb559498224063dc633c3a63f958598108eca29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j24lb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:07:16Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-kl5s9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:20Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:20 crc kubenswrapper[4953]: I1203 00:07:20.683866 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5cmlc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7e11a1d3-ecee-4bf0-87bf-085fab97b73d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d42390b9c53efc63add8ddcebf03c976f5e5448eaa2c34e076b2dc90dbac60fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6x2sc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:07:18Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5cmlc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:20Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:20 crc kubenswrapper[4953]: I1203 00:07:20.686883 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:20 crc kubenswrapper[4953]: I1203 00:07:20.686942 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:20 crc kubenswrapper[4953]: I1203 00:07:20.686959 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:20 crc kubenswrapper[4953]: I1203 00:07:20.687014 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:20 crc kubenswrapper[4953]: I1203 00:07:20.687034 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:20Z","lastTransitionTime":"2025-12-03T00:07:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:20 crc kubenswrapper[4953]: I1203 00:07:20.699418 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cf821840-5353-4620-8f73-99b56b371ab0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cae5e2994fcc24dc9b2ea73e79572615b3311edc42642b7f278fb84123989b23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://29e4b9d1b70a93fb133e0759ad148ae133766a66efeae3ff06a21dfae0e805ff\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dec43536772aa7c3f5a662fa7f3bd73b1cae2791ad454f85792287828d0ffeed\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://034e0bac6892d361700b8f33d6427b8a55e2bd3f89b14ea77a374c6be94d25e2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://761973393a93c59c23e69f41b8ac0d16abc35abd61ecbc6d51a42d94cd664233\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T00:07:09Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 00:07:03.713374 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 00:07:03.715052 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1528563048/tls.crt::/tmp/serving-cert-1528563048/tls.key\\\\\\\"\\\\nI1203 00:07:09.300052 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 00:07:09.304192 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 00:07:09.304255 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 00:07:09.304309 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 00:07:09.304323 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 00:07:09.313563 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 00:07:09.313597 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 00:07:09.313603 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 00:07:09.313609 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 00:07:09.313613 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 00:07:09.313618 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 00:07:09.313622 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 00:07:09.313709 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 00:07:09.327347 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T00:06:53Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c143d4232fb8bf7ef5999a664e431a00ab0db030f4d31a28d0bb21e00e38581\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:53Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72af6c27296eae2e275d455795c68f0d272869d63642b8353b17d466d0c901cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://72af6c27296eae2e275d455795c68f0d272869d63642b8353b17d466d0c901cf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:06:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:06:51Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:20Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:20 crc kubenswrapper[4953]: I1203 00:07:20.715171 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:10Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:20Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:20 crc kubenswrapper[4953]: I1203 00:07:20.727490 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce8b414e1fc22eb30decb6a6411b672e7f443055cb89ebb029fbf6ffc1d5dca7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:20Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:20 crc kubenswrapper[4953]: I1203 00:07:20.738671 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-6dfbn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53ffb1e9-9a2e-4869-80d7-776f4aa7ef8b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ab4205c0d38ad6a0c32077eadb1fd5fd3a279d9c8b8a2ffbe693ec2c4018370c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ct26c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7df84405eda520ad5a22d35750edf07ef3c3977a756ad5739c54f91cba9da7b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ct26c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:07:16Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-6dfbn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:20Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:20 crc kubenswrapper[4953]: I1203 00:07:20.756301 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-txffx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8816ebee-c735-4b8d-b1ec-65349707b347\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:16Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:16Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af4c7764fca87144d01f3738ea010308744ccf50c54b7c471ccc213112de00d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af4c7764fca87144d01f3738ea010308744ccf50c54b7c471ccc213112de00d5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:07:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:07:16Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-txffx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:20Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:20 crc kubenswrapper[4953]: I1203 00:07:20.789264 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:20 crc kubenswrapper[4953]: I1203 00:07:20.789305 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:20 crc kubenswrapper[4953]: I1203 00:07:20.789318 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:20 crc kubenswrapper[4953]: I1203 00:07:20.789334 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:20 crc kubenswrapper[4953]: I1203 00:07:20.789345 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:20Z","lastTransitionTime":"2025-12-03T00:07:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:20 crc kubenswrapper[4953]: I1203 00:07:20.891448 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:20 crc kubenswrapper[4953]: I1203 00:07:20.891490 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:20 crc kubenswrapper[4953]: I1203 00:07:20.891501 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:20 crc kubenswrapper[4953]: I1203 00:07:20.891517 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:20 crc kubenswrapper[4953]: I1203 00:07:20.891529 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:20Z","lastTransitionTime":"2025-12-03T00:07:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:20 crc kubenswrapper[4953]: I1203 00:07:20.994308 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:20 crc kubenswrapper[4953]: I1203 00:07:20.994339 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:20 crc kubenswrapper[4953]: I1203 00:07:20.994348 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:20 crc kubenswrapper[4953]: I1203 00:07:20.994360 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:20 crc kubenswrapper[4953]: I1203 00:07:20.994370 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:20Z","lastTransitionTime":"2025-12-03T00:07:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:21 crc kubenswrapper[4953]: I1203 00:07:21.097259 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:21 crc kubenswrapper[4953]: I1203 00:07:21.097307 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:21 crc kubenswrapper[4953]: I1203 00:07:21.097323 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:21 crc kubenswrapper[4953]: I1203 00:07:21.097345 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:21 crc kubenswrapper[4953]: I1203 00:07:21.097356 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:21Z","lastTransitionTime":"2025-12-03T00:07:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:21 crc kubenswrapper[4953]: I1203 00:07:21.199928 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:21 crc kubenswrapper[4953]: I1203 00:07:21.200006 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:21 crc kubenswrapper[4953]: I1203 00:07:21.200022 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:21 crc kubenswrapper[4953]: I1203 00:07:21.200041 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:21 crc kubenswrapper[4953]: I1203 00:07:21.200057 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:21Z","lastTransitionTime":"2025-12-03T00:07:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:21 crc kubenswrapper[4953]: I1203 00:07:21.303589 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:21 crc kubenswrapper[4953]: I1203 00:07:21.304289 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:21 crc kubenswrapper[4953]: I1203 00:07:21.304313 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:21 crc kubenswrapper[4953]: I1203 00:07:21.304348 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:21 crc kubenswrapper[4953]: I1203 00:07:21.304371 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:21Z","lastTransitionTime":"2025-12-03T00:07:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:21 crc kubenswrapper[4953]: I1203 00:07:21.362389 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:21Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:21 crc kubenswrapper[4953]: I1203 00:07:21.375376 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2ed9c7fa96c8951fbc1728fe7a91fefdd999077285621df9c1abedf9ae0c449\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:21Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:21 crc kubenswrapper[4953]: I1203 00:07:21.393539 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:10Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:21Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:21 crc kubenswrapper[4953]: I1203 00:07:21.407241 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-675qk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"43a4bad8-71f4-441a-8302-aeb526a986af\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5b5794fef6477aab5deed8f642e8cc50b37eb1fd0cc6f19c003f34ad31c1e3c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gdrw4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:07:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-675qk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:21Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:21 crc kubenswrapper[4953]: I1203 00:07:21.407931 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:21 crc kubenswrapper[4953]: I1203 00:07:21.408079 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:21 crc kubenswrapper[4953]: I1203 00:07:21.408167 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:21 crc kubenswrapper[4953]: I1203 00:07:21.408251 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:21 crc kubenswrapper[4953]: I1203 00:07:21.408340 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:21Z","lastTransitionTime":"2025-12-03T00:07:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:21 crc kubenswrapper[4953]: I1203 00:07:21.422895 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f94f8c51-c665-4ca3-bf17-7a04ea5c7ade\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72c5cd7dc1c3a8adfedab6ba43a59d80b3efe65f2bb56be95e76cc86a409dcf3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2810ffc1b141fa7da811406053db9d72c708edf5f2bb1dd75e4d52ab35fd24c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c67b4277b20a9f1287f50ef82d0785c9d02543a8e7ed26e0413eae332b866db7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://258d78e1ad9d3dcda6976909ae3fe71aecead5b995fc66b3fd6e2722d80b5ab8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:06:51Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:21Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:21 crc kubenswrapper[4953]: I1203 00:07:21.437196 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09657a9f448a4c92a759c73a0a892076d7eb46393e6fd565e5f73f34c71aac23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5af5eac4a58870f63497f99b503cfa5fb922278095e0332ec50376925ecf2109\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:21Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:21 crc kubenswrapper[4953]: I1203 00:07:21.457446 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-wj6rz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"96aa67a0-15dd-420d-a1d5-ed239d7df1fb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:16Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:16Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:16Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kf62m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b12bfb5bdf9fd4f2c50a335fc82df6ee0440eaf8741897da03aa2dae368cc6b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b12bfb5bdf9fd4f2c50a335fc82df6ee0440eaf8741897da03aa2dae368cc6b4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:07:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kf62m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83aaa5fa28110bc0ef05aff75cd1442bc9bbdd33b822ae2ef22d4e4ab83238bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://83aaa5fa28110bc0ef05aff75cd1442bc9bbdd33b822ae2ef22d4e4ab83238bc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:07:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:07:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kf62m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ab6213a33a19fedebce6d74a0ebbb3b3a2de5687504e21cc24093cc42ade635\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ab6213a33a19fedebce6d74a0ebbb3b3a2de5687504e21cc24093cc42ade635\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:07:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:07:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kf62m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://de480488843afa6b8dede1e1b42b5181dca1993e2f39dc1b461a18a57e3668e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://de480488843afa6b8dede1e1b42b5181dca1993e2f39dc1b461a18a57e3668e1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:07:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kf62m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kf62m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kf62m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:07:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-wj6rz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:21Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:21 crc kubenswrapper[4953]: I1203 00:07:21.471341 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-kl5s9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d85e5c97-16ad-4bd7-9e73-b7f12678be97\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c21df91a9e4e3165686882b2eb559498224063dc633c3a63f958598108eca29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j24lb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:07:16Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-kl5s9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:21Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:21 crc kubenswrapper[4953]: I1203 00:07:21.491414 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cf821840-5353-4620-8f73-99b56b371ab0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cae5e2994fcc24dc9b2ea73e79572615b3311edc42642b7f278fb84123989b23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://29e4b9d1b70a93fb133e0759ad148ae133766a66efeae3ff06a21dfae0e805ff\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dec43536772aa7c3f5a662fa7f3bd73b1cae2791ad454f85792287828d0ffeed\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://034e0bac6892d361700b8f33d6427b8a55e2bd3f89b14ea77a374c6be94d25e2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://761973393a93c59c23e69f41b8ac0d16abc35abd61ecbc6d51a42d94cd664233\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T00:07:09Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 00:07:03.713374 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 00:07:03.715052 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1528563048/tls.crt::/tmp/serving-cert-1528563048/tls.key\\\\\\\"\\\\nI1203 00:07:09.300052 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 00:07:09.304192 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 00:07:09.304255 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 00:07:09.304309 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 00:07:09.304323 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 00:07:09.313563 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 00:07:09.313597 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 00:07:09.313603 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 00:07:09.313609 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 00:07:09.313613 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 00:07:09.313618 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 00:07:09.313622 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 00:07:09.313709 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 00:07:09.327347 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T00:06:53Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c143d4232fb8bf7ef5999a664e431a00ab0db030f4d31a28d0bb21e00e38581\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:53Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72af6c27296eae2e275d455795c68f0d272869d63642b8353b17d466d0c901cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://72af6c27296eae2e275d455795c68f0d272869d63642b8353b17d466d0c901cf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:06:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:06:51Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:21Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:21 crc kubenswrapper[4953]: I1203 00:07:21.506907 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:10Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:21Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:21 crc kubenswrapper[4953]: I1203 00:07:21.512411 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:21 crc kubenswrapper[4953]: I1203 00:07:21.512460 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:21 crc kubenswrapper[4953]: I1203 00:07:21.512483 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:21 crc kubenswrapper[4953]: I1203 00:07:21.512507 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:21 crc kubenswrapper[4953]: I1203 00:07:21.512526 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:21Z","lastTransitionTime":"2025-12-03T00:07:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:21 crc kubenswrapper[4953]: I1203 00:07:21.524228 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce8b414e1fc22eb30decb6a6411b672e7f443055cb89ebb029fbf6ffc1d5dca7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:21Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:21 crc kubenswrapper[4953]: I1203 00:07:21.538185 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-6dfbn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53ffb1e9-9a2e-4869-80d7-776f4aa7ef8b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ab4205c0d38ad6a0c32077eadb1fd5fd3a279d9c8b8a2ffbe693ec2c4018370c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ct26c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7df84405eda520ad5a22d35750edf07ef3c3977a756ad5739c54f91cba9da7b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ct26c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:07:16Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-6dfbn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:21Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:21 crc kubenswrapper[4953]: I1203 00:07:21.555145 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-txffx" event={"ID":"8816ebee-c735-4b8d-b1ec-65349707b347","Type":"ContainerStarted","Data":"7deadc8e135ff1bc7c5a2f093e489bb5b4b6f166bc874c6acc2a841c5397f11e"} Dec 03 00:07:21 crc kubenswrapper[4953]: I1203 00:07:21.559899 4953 generic.go:334] "Generic (PLEG): container finished" podID="96aa67a0-15dd-420d-a1d5-ed239d7df1fb" containerID="97be4ba692bc61078d4b0ba39fb0c12772548c6c9251ff3356ea702f151223cd" exitCode=0 Dec 03 00:07:21 crc kubenswrapper[4953]: I1203 00:07:21.560054 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-wj6rz" event={"ID":"96aa67a0-15dd-420d-a1d5-ed239d7df1fb","Type":"ContainerDied","Data":"97be4ba692bc61078d4b0ba39fb0c12772548c6c9251ff3356ea702f151223cd"} Dec 03 00:07:21 crc kubenswrapper[4953]: I1203 00:07:21.562120 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-txffx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8816ebee-c735-4b8d-b1ec-65349707b347\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:16Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:16Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af4c7764fca87144d01f3738ea010308744ccf50c54b7c471ccc213112de00d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af4c7764fca87144d01f3738ea010308744ccf50c54b7c471ccc213112de00d5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:07:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:07:16Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-txffx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:21Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:21 crc kubenswrapper[4953]: I1203 00:07:21.576576 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5cmlc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7e11a1d3-ecee-4bf0-87bf-085fab97b73d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d42390b9c53efc63add8ddcebf03c976f5e5448eaa2c34e076b2dc90dbac60fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6x2sc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:07:18Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5cmlc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:21Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:21 crc kubenswrapper[4953]: I1203 00:07:21.589148 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09657a9f448a4c92a759c73a0a892076d7eb46393e6fd565e5f73f34c71aac23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5af5eac4a58870f63497f99b503cfa5fb922278095e0332ec50376925ecf2109\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:21Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:21 crc kubenswrapper[4953]: I1203 00:07:21.605150 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-wj6rz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"96aa67a0-15dd-420d-a1d5-ed239d7df1fb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:16Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:16Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:16Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kf62m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b12bfb5bdf9fd4f2c50a335fc82df6ee0440eaf8741897da03aa2dae368cc6b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b12bfb5bdf9fd4f2c50a335fc82df6ee0440eaf8741897da03aa2dae368cc6b4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:07:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kf62m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83aaa5fa28110bc0ef05aff75cd1442bc9bbdd33b822ae2ef22d4e4ab83238bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://83aaa5fa28110bc0ef05aff75cd1442bc9bbdd33b822ae2ef22d4e4ab83238bc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:07:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:07:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kf62m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ab6213a33a19fedebce6d74a0ebbb3b3a2de5687504e21cc24093cc42ade635\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ab6213a33a19fedebce6d74a0ebbb3b3a2de5687504e21cc24093cc42ade635\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:07:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:07:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kf62m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://de480488843afa6b8dede1e1b42b5181dca1993e2f39dc1b461a18a57e3668e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://de480488843afa6b8dede1e1b42b5181dca1993e2f39dc1b461a18a57e3668e1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:07:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kf62m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://97be4ba692bc61078d4b0ba39fb0c12772548c6c9251ff3356ea702f151223cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://97be4ba692bc61078d4b0ba39fb0c12772548c6c9251ff3356ea702f151223cd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:07:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kf62m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kf62m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:07:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-wj6rz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:21Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:21 crc kubenswrapper[4953]: I1203 00:07:21.618224 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-kl5s9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d85e5c97-16ad-4bd7-9e73-b7f12678be97\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c21df91a9e4e3165686882b2eb559498224063dc633c3a63f958598108eca29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j24lb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:07:16Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-kl5s9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:21Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:21 crc kubenswrapper[4953]: I1203 00:07:21.635380 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:21 crc kubenswrapper[4953]: I1203 00:07:21.635564 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:21 crc kubenswrapper[4953]: I1203 00:07:21.635673 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:21 crc kubenswrapper[4953]: I1203 00:07:21.635694 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:21 crc kubenswrapper[4953]: I1203 00:07:21.635711 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:21Z","lastTransitionTime":"2025-12-03T00:07:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:21 crc kubenswrapper[4953]: I1203 00:07:21.644718 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f94f8c51-c665-4ca3-bf17-7a04ea5c7ade\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72c5cd7dc1c3a8adfedab6ba43a59d80b3efe65f2bb56be95e76cc86a409dcf3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2810ffc1b141fa7da811406053db9d72c708edf5f2bb1dd75e4d52ab35fd24c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c67b4277b20a9f1287f50ef82d0785c9d02543a8e7ed26e0413eae332b866db7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://258d78e1ad9d3dcda6976909ae3fe71aecead5b995fc66b3fd6e2722d80b5ab8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:06:51Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:21Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:21 crc kubenswrapper[4953]: I1203 00:07:21.657092 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce8b414e1fc22eb30decb6a6411b672e7f443055cb89ebb029fbf6ffc1d5dca7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:21Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:21 crc kubenswrapper[4953]: I1203 00:07:21.669641 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-6dfbn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53ffb1e9-9a2e-4869-80d7-776f4aa7ef8b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ab4205c0d38ad6a0c32077eadb1fd5fd3a279d9c8b8a2ffbe693ec2c4018370c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ct26c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7df84405eda520ad5a22d35750edf07ef3c3977a756ad5739c54f91cba9da7b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ct26c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:07:16Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-6dfbn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:21Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:21 crc kubenswrapper[4953]: I1203 00:07:21.703772 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-txffx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8816ebee-c735-4b8d-b1ec-65349707b347\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:16Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:16Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af4c7764fca87144d01f3738ea010308744ccf50c54b7c471ccc213112de00d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af4c7764fca87144d01f3738ea010308744ccf50c54b7c471ccc213112de00d5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:07:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:07:16Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-txffx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:21Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:21 crc kubenswrapper[4953]: I1203 00:07:21.715681 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5cmlc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7e11a1d3-ecee-4bf0-87bf-085fab97b73d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d42390b9c53efc63add8ddcebf03c976f5e5448eaa2c34e076b2dc90dbac60fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6x2sc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:07:18Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5cmlc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:21Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:21 crc kubenswrapper[4953]: I1203 00:07:21.728933 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cf821840-5353-4620-8f73-99b56b371ab0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cae5e2994fcc24dc9b2ea73e79572615b3311edc42642b7f278fb84123989b23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://29e4b9d1b70a93fb133e0759ad148ae133766a66efeae3ff06a21dfae0e805ff\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dec43536772aa7c3f5a662fa7f3bd73b1cae2791ad454f85792287828d0ffeed\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://034e0bac6892d361700b8f33d6427b8a55e2bd3f89b14ea77a374c6be94d25e2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://761973393a93c59c23e69f41b8ac0d16abc35abd61ecbc6d51a42d94cd664233\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T00:07:09Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 00:07:03.713374 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 00:07:03.715052 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1528563048/tls.crt::/tmp/serving-cert-1528563048/tls.key\\\\\\\"\\\\nI1203 00:07:09.300052 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 00:07:09.304192 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 00:07:09.304255 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 00:07:09.304309 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 00:07:09.304323 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 00:07:09.313563 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 00:07:09.313597 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 00:07:09.313603 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 00:07:09.313609 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 00:07:09.313613 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 00:07:09.313618 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 00:07:09.313622 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 00:07:09.313709 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 00:07:09.327347 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T00:06:53Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c143d4232fb8bf7ef5999a664e431a00ab0db030f4d31a28d0bb21e00e38581\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:53Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72af6c27296eae2e275d455795c68f0d272869d63642b8353b17d466d0c901cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://72af6c27296eae2e275d455795c68f0d272869d63642b8353b17d466d0c901cf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:06:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:06:51Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:21Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:21 crc kubenswrapper[4953]: I1203 00:07:21.737832 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:21 crc kubenswrapper[4953]: I1203 00:07:21.738182 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:21 crc kubenswrapper[4953]: I1203 00:07:21.738269 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:21 crc kubenswrapper[4953]: I1203 00:07:21.738361 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:21 crc kubenswrapper[4953]: I1203 00:07:21.738449 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:21Z","lastTransitionTime":"2025-12-03T00:07:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:21 crc kubenswrapper[4953]: I1203 00:07:21.741773 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:10Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:21Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:21 crc kubenswrapper[4953]: I1203 00:07:21.753997 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:21Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:21 crc kubenswrapper[4953]: I1203 00:07:21.766344 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2ed9c7fa96c8951fbc1728fe7a91fefdd999077285621df9c1abedf9ae0c449\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:21Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:21 crc kubenswrapper[4953]: I1203 00:07:21.779483 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:10Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:21Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:21 crc kubenswrapper[4953]: I1203 00:07:21.794216 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-675qk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"43a4bad8-71f4-441a-8302-aeb526a986af\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5b5794fef6477aab5deed8f642e8cc50b37eb1fd0cc6f19c003f34ad31c1e3c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gdrw4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:07:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-675qk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:21Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:21 crc kubenswrapper[4953]: I1203 00:07:21.840446 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:21 crc kubenswrapper[4953]: I1203 00:07:21.840475 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:21 crc kubenswrapper[4953]: I1203 00:07:21.840483 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:21 crc kubenswrapper[4953]: I1203 00:07:21.840495 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:21 crc kubenswrapper[4953]: I1203 00:07:21.840505 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:21Z","lastTransitionTime":"2025-12-03T00:07:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:21 crc kubenswrapper[4953]: I1203 00:07:21.942867 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:21 crc kubenswrapper[4953]: I1203 00:07:21.943086 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:21 crc kubenswrapper[4953]: I1203 00:07:21.943410 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:21 crc kubenswrapper[4953]: I1203 00:07:21.943562 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:21 crc kubenswrapper[4953]: I1203 00:07:21.943629 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:21Z","lastTransitionTime":"2025-12-03T00:07:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:22 crc kubenswrapper[4953]: I1203 00:07:22.046761 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:22 crc kubenswrapper[4953]: I1203 00:07:22.047067 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:22 crc kubenswrapper[4953]: I1203 00:07:22.047203 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:22 crc kubenswrapper[4953]: I1203 00:07:22.047311 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:22 crc kubenswrapper[4953]: I1203 00:07:22.047391 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:22Z","lastTransitionTime":"2025-12-03T00:07:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:22 crc kubenswrapper[4953]: I1203 00:07:22.150295 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:22 crc kubenswrapper[4953]: I1203 00:07:22.150778 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:22 crc kubenswrapper[4953]: I1203 00:07:22.150850 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:22 crc kubenswrapper[4953]: I1203 00:07:22.150931 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:22 crc kubenswrapper[4953]: I1203 00:07:22.151047 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:22Z","lastTransitionTime":"2025-12-03T00:07:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:22 crc kubenswrapper[4953]: I1203 00:07:22.253213 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:22 crc kubenswrapper[4953]: I1203 00:07:22.253465 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:22 crc kubenswrapper[4953]: I1203 00:07:22.253525 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:22 crc kubenswrapper[4953]: I1203 00:07:22.253608 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:22 crc kubenswrapper[4953]: I1203 00:07:22.253680 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:22Z","lastTransitionTime":"2025-12-03T00:07:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:22 crc kubenswrapper[4953]: I1203 00:07:22.347074 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 00:07:22 crc kubenswrapper[4953]: I1203 00:07:22.347268 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 00:07:22 crc kubenswrapper[4953]: E1203 00:07:22.347605 4953 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 00:07:22 crc kubenswrapper[4953]: E1203 00:07:22.347629 4953 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 00:07:22 crc kubenswrapper[4953]: I1203 00:07:22.347306 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 00:07:22 crc kubenswrapper[4953]: E1203 00:07:22.348349 4953 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 00:07:22 crc kubenswrapper[4953]: I1203 00:07:22.356286 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:22 crc kubenswrapper[4953]: I1203 00:07:22.356568 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:22 crc kubenswrapper[4953]: I1203 00:07:22.356697 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:22 crc kubenswrapper[4953]: I1203 00:07:22.356825 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:22 crc kubenswrapper[4953]: I1203 00:07:22.357031 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:22Z","lastTransitionTime":"2025-12-03T00:07:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:22 crc kubenswrapper[4953]: I1203 00:07:22.460467 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:22 crc kubenswrapper[4953]: I1203 00:07:22.460563 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:22 crc kubenswrapper[4953]: I1203 00:07:22.460589 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:22 crc kubenswrapper[4953]: I1203 00:07:22.460618 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:22 crc kubenswrapper[4953]: I1203 00:07:22.460645 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:22Z","lastTransitionTime":"2025-12-03T00:07:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:22 crc kubenswrapper[4953]: I1203 00:07:22.562700 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:22 crc kubenswrapper[4953]: I1203 00:07:22.562732 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:22 crc kubenswrapper[4953]: I1203 00:07:22.562740 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:22 crc kubenswrapper[4953]: I1203 00:07:22.562755 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:22 crc kubenswrapper[4953]: I1203 00:07:22.562764 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:22Z","lastTransitionTime":"2025-12-03T00:07:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:22 crc kubenswrapper[4953]: I1203 00:07:22.566432 4953 generic.go:334] "Generic (PLEG): container finished" podID="96aa67a0-15dd-420d-a1d5-ed239d7df1fb" containerID="b1a1f009fcbd8af9217db1dc97f662719f8c1db232df0e2804e44c881abb1df2" exitCode=0 Dec 03 00:07:22 crc kubenswrapper[4953]: I1203 00:07:22.566463 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-wj6rz" event={"ID":"96aa67a0-15dd-420d-a1d5-ed239d7df1fb","Type":"ContainerDied","Data":"b1a1f009fcbd8af9217db1dc97f662719f8c1db232df0e2804e44c881abb1df2"} Dec 03 00:07:22 crc kubenswrapper[4953]: I1203 00:07:22.581852 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:10Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:22Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:22 crc kubenswrapper[4953]: I1203 00:07:22.597235 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-675qk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"43a4bad8-71f4-441a-8302-aeb526a986af\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5b5794fef6477aab5deed8f642e8cc50b37eb1fd0cc6f19c003f34ad31c1e3c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gdrw4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:07:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-675qk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:22Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:22 crc kubenswrapper[4953]: I1203 00:07:22.615324 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f94f8c51-c665-4ca3-bf17-7a04ea5c7ade\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72c5cd7dc1c3a8adfedab6ba43a59d80b3efe65f2bb56be95e76cc86a409dcf3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2810ffc1b141fa7da811406053db9d72c708edf5f2bb1dd75e4d52ab35fd24c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c67b4277b20a9f1287f50ef82d0785c9d02543a8e7ed26e0413eae332b866db7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://258d78e1ad9d3dcda6976909ae3fe71aecead5b995fc66b3fd6e2722d80b5ab8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:06:51Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:22Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:22 crc kubenswrapper[4953]: I1203 00:07:22.635756 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09657a9f448a4c92a759c73a0a892076d7eb46393e6fd565e5f73f34c71aac23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5af5eac4a58870f63497f99b503cfa5fb922278095e0332ec50376925ecf2109\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:22Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:22 crc kubenswrapper[4953]: I1203 00:07:22.655481 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-wj6rz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"96aa67a0-15dd-420d-a1d5-ed239d7df1fb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:16Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:16Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kf62m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b12bfb5bdf9fd4f2c50a335fc82df6ee0440eaf8741897da03aa2dae368cc6b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b12bfb5bdf9fd4f2c50a335fc82df6ee0440eaf8741897da03aa2dae368cc6b4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:07:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kf62m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83aaa5fa28110bc0ef05aff75cd1442bc9bbdd33b822ae2ef22d4e4ab83238bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://83aaa5fa28110bc0ef05aff75cd1442bc9bbdd33b822ae2ef22d4e4ab83238bc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:07:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:07:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kf62m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ab6213a33a19fedebce6d74a0ebbb3b3a2de5687504e21cc24093cc42ade635\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ab6213a33a19fedebce6d74a0ebbb3b3a2de5687504e21cc24093cc42ade635\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:07:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:07:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kf62m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://de480488843afa6b8dede1e1b42b5181dca1993e2f39dc1b461a18a57e3668e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://de480488843afa6b8dede1e1b42b5181dca1993e2f39dc1b461a18a57e3668e1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:07:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kf62m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://97be4ba692bc61078d4b0ba39fb0c12772548c6c9251ff3356ea702f151223cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://97be4ba692bc61078d4b0ba39fb0c12772548c6c9251ff3356ea702f151223cd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:07:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kf62m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1a1f009fcbd8af9217db1dc97f662719f8c1db232df0e2804e44c881abb1df2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b1a1f009fcbd8af9217db1dc97f662719f8c1db232df0e2804e44c881abb1df2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:07:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:07:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kf62m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:07:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-wj6rz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:22Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:22 crc kubenswrapper[4953]: I1203 00:07:22.666199 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:22 crc kubenswrapper[4953]: I1203 00:07:22.666278 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:22 crc kubenswrapper[4953]: I1203 00:07:22.666306 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:22 crc kubenswrapper[4953]: I1203 00:07:22.666339 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:22 crc kubenswrapper[4953]: I1203 00:07:22.666362 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:22Z","lastTransitionTime":"2025-12-03T00:07:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:22 crc kubenswrapper[4953]: I1203 00:07:22.670253 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-kl5s9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d85e5c97-16ad-4bd7-9e73-b7f12678be97\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c21df91a9e4e3165686882b2eb559498224063dc633c3a63f958598108eca29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j24lb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:07:16Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-kl5s9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:22Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:22 crc kubenswrapper[4953]: I1203 00:07:22.690637 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cf821840-5353-4620-8f73-99b56b371ab0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cae5e2994fcc24dc9b2ea73e79572615b3311edc42642b7f278fb84123989b23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://29e4b9d1b70a93fb133e0759ad148ae133766a66efeae3ff06a21dfae0e805ff\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dec43536772aa7c3f5a662fa7f3bd73b1cae2791ad454f85792287828d0ffeed\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://034e0bac6892d361700b8f33d6427b8a55e2bd3f89b14ea77a374c6be94d25e2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://761973393a93c59c23e69f41b8ac0d16abc35abd61ecbc6d51a42d94cd664233\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T00:07:09Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 00:07:03.713374 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 00:07:03.715052 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1528563048/tls.crt::/tmp/serving-cert-1528563048/tls.key\\\\\\\"\\\\nI1203 00:07:09.300052 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 00:07:09.304192 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 00:07:09.304255 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 00:07:09.304309 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 00:07:09.304323 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 00:07:09.313563 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 00:07:09.313597 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 00:07:09.313603 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 00:07:09.313609 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 00:07:09.313613 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 00:07:09.313618 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 00:07:09.313622 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 00:07:09.313709 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 00:07:09.327347 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T00:06:53Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c143d4232fb8bf7ef5999a664e431a00ab0db030f4d31a28d0bb21e00e38581\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:53Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72af6c27296eae2e275d455795c68f0d272869d63642b8353b17d466d0c901cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://72af6c27296eae2e275d455795c68f0d272869d63642b8353b17d466d0c901cf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:06:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:06:51Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:22Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:22 crc kubenswrapper[4953]: I1203 00:07:22.704958 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:10Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:22Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:22 crc kubenswrapper[4953]: I1203 00:07:22.721130 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce8b414e1fc22eb30decb6a6411b672e7f443055cb89ebb029fbf6ffc1d5dca7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:22Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:22 crc kubenswrapper[4953]: I1203 00:07:22.733524 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-6dfbn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53ffb1e9-9a2e-4869-80d7-776f4aa7ef8b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ab4205c0d38ad6a0c32077eadb1fd5fd3a279d9c8b8a2ffbe693ec2c4018370c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ct26c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7df84405eda520ad5a22d35750edf07ef3c3977a756ad5739c54f91cba9da7b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ct26c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:07:16Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-6dfbn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:22Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:22 crc kubenswrapper[4953]: I1203 00:07:22.769795 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:22 crc kubenswrapper[4953]: I1203 00:07:22.769944 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:22 crc kubenswrapper[4953]: I1203 00:07:22.770030 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:22 crc kubenswrapper[4953]: I1203 00:07:22.770057 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:22 crc kubenswrapper[4953]: I1203 00:07:22.770110 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:22Z","lastTransitionTime":"2025-12-03T00:07:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:22 crc kubenswrapper[4953]: I1203 00:07:22.772319 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-txffx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8816ebee-c735-4b8d-b1ec-65349707b347\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:16Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:16Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af4c7764fca87144d01f3738ea010308744ccf50c54b7c471ccc213112de00d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af4c7764fca87144d01f3738ea010308744ccf50c54b7c471ccc213112de00d5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:07:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:07:16Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-txffx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:22Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:22 crc kubenswrapper[4953]: I1203 00:07:22.785463 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5cmlc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7e11a1d3-ecee-4bf0-87bf-085fab97b73d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d42390b9c53efc63add8ddcebf03c976f5e5448eaa2c34e076b2dc90dbac60fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6x2sc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:07:18Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5cmlc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:22Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:22 crc kubenswrapper[4953]: I1203 00:07:22.800866 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:22Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:22 crc kubenswrapper[4953]: I1203 00:07:22.811927 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2ed9c7fa96c8951fbc1728fe7a91fefdd999077285621df9c1abedf9ae0c449\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:22Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:22 crc kubenswrapper[4953]: I1203 00:07:22.873032 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:22 crc kubenswrapper[4953]: I1203 00:07:22.873096 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:22 crc kubenswrapper[4953]: I1203 00:07:22.873115 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:22 crc kubenswrapper[4953]: I1203 00:07:22.873140 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:22 crc kubenswrapper[4953]: I1203 00:07:22.873159 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:22Z","lastTransitionTime":"2025-12-03T00:07:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:22 crc kubenswrapper[4953]: I1203 00:07:22.901337 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:22 crc kubenswrapper[4953]: I1203 00:07:22.901463 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:22 crc kubenswrapper[4953]: I1203 00:07:22.901531 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:22 crc kubenswrapper[4953]: I1203 00:07:22.901590 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:22 crc kubenswrapper[4953]: I1203 00:07:22.901670 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:22Z","lastTransitionTime":"2025-12-03T00:07:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:22 crc kubenswrapper[4953]: E1203 00:07:22.919748 4953 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:07:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:22Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:07:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:22Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:07:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:22Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:07:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:22Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"eb4ae5a6-0f6f-49ef-a027-1cdee9d3c536\\\",\\\"systemUUID\\\":\\\"4c1f3d44-3d3d-425d-993f-471223206a89\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:22Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:22 crc kubenswrapper[4953]: I1203 00:07:22.923270 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:22 crc kubenswrapper[4953]: I1203 00:07:22.923296 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:22 crc kubenswrapper[4953]: I1203 00:07:22.923338 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:22 crc kubenswrapper[4953]: I1203 00:07:22.923352 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:22 crc kubenswrapper[4953]: I1203 00:07:22.923360 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:22Z","lastTransitionTime":"2025-12-03T00:07:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:22 crc kubenswrapper[4953]: E1203 00:07:22.935414 4953 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:07:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:22Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:07:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:22Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:07:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:22Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:07:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:22Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"eb4ae5a6-0f6f-49ef-a027-1cdee9d3c536\\\",\\\"systemUUID\\\":\\\"4c1f3d44-3d3d-425d-993f-471223206a89\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:22Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:22 crc kubenswrapper[4953]: I1203 00:07:22.938408 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:22 crc kubenswrapper[4953]: I1203 00:07:22.938619 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:22 crc kubenswrapper[4953]: I1203 00:07:22.938634 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:22 crc kubenswrapper[4953]: I1203 00:07:22.938996 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:22 crc kubenswrapper[4953]: I1203 00:07:22.939024 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:22Z","lastTransitionTime":"2025-12-03T00:07:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:22 crc kubenswrapper[4953]: E1203 00:07:22.955419 4953 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:07:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:22Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:07:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:22Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:07:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:22Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:07:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:22Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"eb4ae5a6-0f6f-49ef-a027-1cdee9d3c536\\\",\\\"systemUUID\\\":\\\"4c1f3d44-3d3d-425d-993f-471223206a89\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:22Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:22 crc kubenswrapper[4953]: I1203 00:07:22.959749 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:22 crc kubenswrapper[4953]: I1203 00:07:22.959776 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:22 crc kubenswrapper[4953]: I1203 00:07:22.959784 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:22 crc kubenswrapper[4953]: I1203 00:07:22.959796 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:22 crc kubenswrapper[4953]: I1203 00:07:22.959805 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:22Z","lastTransitionTime":"2025-12-03T00:07:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:22 crc kubenswrapper[4953]: E1203 00:07:22.971325 4953 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:07:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:22Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:07:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:22Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:07:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:22Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:07:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:22Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"eb4ae5a6-0f6f-49ef-a027-1cdee9d3c536\\\",\\\"systemUUID\\\":\\\"4c1f3d44-3d3d-425d-993f-471223206a89\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:22Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:22 crc kubenswrapper[4953]: I1203 00:07:22.974428 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:22 crc kubenswrapper[4953]: I1203 00:07:22.974455 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:22 crc kubenswrapper[4953]: I1203 00:07:22.974462 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:22 crc kubenswrapper[4953]: I1203 00:07:22.974474 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:22 crc kubenswrapper[4953]: I1203 00:07:22.974482 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:22Z","lastTransitionTime":"2025-12-03T00:07:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:22 crc kubenswrapper[4953]: E1203 00:07:22.987535 4953 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:07:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:22Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:07:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:22Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:07:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:22Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:07:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:22Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"eb4ae5a6-0f6f-49ef-a027-1cdee9d3c536\\\",\\\"systemUUID\\\":\\\"4c1f3d44-3d3d-425d-993f-471223206a89\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:22Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:22 crc kubenswrapper[4953]: E1203 00:07:22.987666 4953 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 03 00:07:22 crc kubenswrapper[4953]: I1203 00:07:22.989025 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:22 crc kubenswrapper[4953]: I1203 00:07:22.989050 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:22 crc kubenswrapper[4953]: I1203 00:07:22.989062 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:22 crc kubenswrapper[4953]: I1203 00:07:22.989078 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:22 crc kubenswrapper[4953]: I1203 00:07:22.989089 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:22Z","lastTransitionTime":"2025-12-03T00:07:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:23 crc kubenswrapper[4953]: I1203 00:07:23.091866 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:23 crc kubenswrapper[4953]: I1203 00:07:23.091922 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:23 crc kubenswrapper[4953]: I1203 00:07:23.091943 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:23 crc kubenswrapper[4953]: I1203 00:07:23.092009 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:23 crc kubenswrapper[4953]: I1203 00:07:23.092035 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:23Z","lastTransitionTime":"2025-12-03T00:07:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:23 crc kubenswrapper[4953]: I1203 00:07:23.198584 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:23 crc kubenswrapper[4953]: I1203 00:07:23.198670 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:23 crc kubenswrapper[4953]: I1203 00:07:23.198695 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:23 crc kubenswrapper[4953]: I1203 00:07:23.198734 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:23 crc kubenswrapper[4953]: I1203 00:07:23.198758 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:23Z","lastTransitionTime":"2025-12-03T00:07:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:23 crc kubenswrapper[4953]: I1203 00:07:23.301452 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:23 crc kubenswrapper[4953]: I1203 00:07:23.301526 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:23 crc kubenswrapper[4953]: I1203 00:07:23.301551 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:23 crc kubenswrapper[4953]: I1203 00:07:23.301583 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:23 crc kubenswrapper[4953]: I1203 00:07:23.301607 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:23Z","lastTransitionTime":"2025-12-03T00:07:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:23 crc kubenswrapper[4953]: I1203 00:07:23.404780 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:23 crc kubenswrapper[4953]: I1203 00:07:23.404818 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:23 crc kubenswrapper[4953]: I1203 00:07:23.404827 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:23 crc kubenswrapper[4953]: I1203 00:07:23.404843 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:23 crc kubenswrapper[4953]: I1203 00:07:23.404853 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:23Z","lastTransitionTime":"2025-12-03T00:07:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:23 crc kubenswrapper[4953]: I1203 00:07:23.506476 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:23 crc kubenswrapper[4953]: I1203 00:07:23.506532 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:23 crc kubenswrapper[4953]: I1203 00:07:23.506552 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:23 crc kubenswrapper[4953]: I1203 00:07:23.506587 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:23 crc kubenswrapper[4953]: I1203 00:07:23.506609 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:23Z","lastTransitionTime":"2025-12-03T00:07:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:23 crc kubenswrapper[4953]: I1203 00:07:23.609532 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:23 crc kubenswrapper[4953]: I1203 00:07:23.609857 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:23 crc kubenswrapper[4953]: I1203 00:07:23.609871 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:23 crc kubenswrapper[4953]: I1203 00:07:23.609888 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:23 crc kubenswrapper[4953]: I1203 00:07:23.609901 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:23Z","lastTransitionTime":"2025-12-03T00:07:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:23 crc kubenswrapper[4953]: I1203 00:07:23.612885 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-txffx" event={"ID":"8816ebee-c735-4b8d-b1ec-65349707b347","Type":"ContainerStarted","Data":"aecc0ae6dd34893c9023da3b94ca01ae96ac1e7cd7f5fd18748cce6c273a7f9e"} Dec 03 00:07:23 crc kubenswrapper[4953]: I1203 00:07:23.613166 4953 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-txffx" Dec 03 00:07:23 crc kubenswrapper[4953]: I1203 00:07:23.619127 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-wj6rz" event={"ID":"96aa67a0-15dd-420d-a1d5-ed239d7df1fb","Type":"ContainerStarted","Data":"01815cfd3f58c3f3f5f2a894c1790e0b69c5060c48dc0b1deff050b6ad7af916"} Dec 03 00:07:23 crc kubenswrapper[4953]: I1203 00:07:23.626799 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2ed9c7fa96c8951fbc1728fe7a91fefdd999077285621df9c1abedf9ae0c449\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:23Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:23 crc kubenswrapper[4953]: I1203 00:07:23.640118 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:23Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:23 crc kubenswrapper[4953]: I1203 00:07:23.650942 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:10Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:23Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:23 crc kubenswrapper[4953]: I1203 00:07:23.663592 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-675qk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"43a4bad8-71f4-441a-8302-aeb526a986af\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5b5794fef6477aab5deed8f642e8cc50b37eb1fd0cc6f19c003f34ad31c1e3c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gdrw4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:07:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-675qk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:23Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:23 crc kubenswrapper[4953]: I1203 00:07:23.675285 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-wj6rz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"96aa67a0-15dd-420d-a1d5-ed239d7df1fb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:16Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:16Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kf62m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b12bfb5bdf9fd4f2c50a335fc82df6ee0440eaf8741897da03aa2dae368cc6b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b12bfb5bdf9fd4f2c50a335fc82df6ee0440eaf8741897da03aa2dae368cc6b4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:07:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kf62m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83aaa5fa28110bc0ef05aff75cd1442bc9bbdd33b822ae2ef22d4e4ab83238bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://83aaa5fa28110bc0ef05aff75cd1442bc9bbdd33b822ae2ef22d4e4ab83238bc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:07:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:07:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kf62m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ab6213a33a19fedebce6d74a0ebbb3b3a2de5687504e21cc24093cc42ade635\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ab6213a33a19fedebce6d74a0ebbb3b3a2de5687504e21cc24093cc42ade635\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:07:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:07:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kf62m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://de480488843afa6b8dede1e1b42b5181dca1993e2f39dc1b461a18a57e3668e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://de480488843afa6b8dede1e1b42b5181dca1993e2f39dc1b461a18a57e3668e1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:07:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kf62m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://97be4ba692bc61078d4b0ba39fb0c12772548c6c9251ff3356ea702f151223cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://97be4ba692bc61078d4b0ba39fb0c12772548c6c9251ff3356ea702f151223cd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:07:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kf62m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1a1f009fcbd8af9217db1dc97f662719f8c1db232df0e2804e44c881abb1df2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b1a1f009fcbd8af9217db1dc97f662719f8c1db232df0e2804e44c881abb1df2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:07:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:07:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kf62m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:07:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-wj6rz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:23Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:23 crc kubenswrapper[4953]: I1203 00:07:23.688306 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-kl5s9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d85e5c97-16ad-4bd7-9e73-b7f12678be97\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c21df91a9e4e3165686882b2eb559498224063dc633c3a63f958598108eca29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j24lb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:07:16Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-kl5s9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:23Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:23 crc kubenswrapper[4953]: I1203 00:07:23.702671 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f94f8c51-c665-4ca3-bf17-7a04ea5c7ade\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72c5cd7dc1c3a8adfedab6ba43a59d80b3efe65f2bb56be95e76cc86a409dcf3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2810ffc1b141fa7da811406053db9d72c708edf5f2bb1dd75e4d52ab35fd24c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c67b4277b20a9f1287f50ef82d0785c9d02543a8e7ed26e0413eae332b866db7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://258d78e1ad9d3dcda6976909ae3fe71aecead5b995fc66b3fd6e2722d80b5ab8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:06:51Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:23Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:23 crc kubenswrapper[4953]: I1203 00:07:23.713311 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:23 crc kubenswrapper[4953]: I1203 00:07:23.713348 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:23 crc kubenswrapper[4953]: I1203 00:07:23.713358 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:23 crc kubenswrapper[4953]: I1203 00:07:23.713376 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:23 crc kubenswrapper[4953]: I1203 00:07:23.713393 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:23Z","lastTransitionTime":"2025-12-03T00:07:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:23 crc kubenswrapper[4953]: I1203 00:07:23.717402 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09657a9f448a4c92a759c73a0a892076d7eb46393e6fd565e5f73f34c71aac23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5af5eac4a58870f63497f99b503cfa5fb922278095e0332ec50376925ecf2109\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:23Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:23 crc kubenswrapper[4953]: I1203 00:07:23.727237 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-6dfbn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53ffb1e9-9a2e-4869-80d7-776f4aa7ef8b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ab4205c0d38ad6a0c32077eadb1fd5fd3a279d9c8b8a2ffbe693ec2c4018370c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ct26c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7df84405eda520ad5a22d35750edf07ef3c3977a756ad5739c54f91cba9da7b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ct26c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:07:16Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-6dfbn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:23Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:23 crc kubenswrapper[4953]: I1203 00:07:23.746541 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-txffx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8816ebee-c735-4b8d-b1ec-65349707b347\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:16Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:16Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4b85224419a4075acb340e4b11095f2adb5c316e285afa4e19295c2181c0d24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0f6f645a58993f052a6f365bb808f62d99f7fea74886c5cc022b2d86112321a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a61b951a9cbab2c48c0f2df8351f97d7ff71bf3674f641fe5eac30155232b224\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd5f428de229749485267e81d34e31b708728b52474032e88460e54747f21e11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://00a2ea4a9bb870fce30e7fa845f787f184e7a19e42c95511655ea8c5f84a26d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc0a5d4cc9d614d8492626a1b114bdc52f510eb5965d842c02c083f08d4849a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aecc0ae6dd34893c9023da3b94ca01ae96ac1e7cd7f5fd18748cce6c273a7f9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7deadc8e135ff1bc7c5a2f093e489bb5b4b6f166bc874c6acc2a841c5397f11e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af4c7764fca87144d01f3738ea010308744ccf50c54b7c471ccc213112de00d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af4c7764fca87144d01f3738ea010308744ccf50c54b7c471ccc213112de00d5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:07:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:07:16Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-txffx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:23Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:23 crc kubenswrapper[4953]: I1203 00:07:23.757666 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5cmlc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7e11a1d3-ecee-4bf0-87bf-085fab97b73d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d42390b9c53efc63add8ddcebf03c976f5e5448eaa2c34e076b2dc90dbac60fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6x2sc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:07:18Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5cmlc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:23Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:23 crc kubenswrapper[4953]: I1203 00:07:23.772868 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cf821840-5353-4620-8f73-99b56b371ab0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cae5e2994fcc24dc9b2ea73e79572615b3311edc42642b7f278fb84123989b23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://29e4b9d1b70a93fb133e0759ad148ae133766a66efeae3ff06a21dfae0e805ff\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dec43536772aa7c3f5a662fa7f3bd73b1cae2791ad454f85792287828d0ffeed\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://034e0bac6892d361700b8f33d6427b8a55e2bd3f89b14ea77a374c6be94d25e2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://761973393a93c59c23e69f41b8ac0d16abc35abd61ecbc6d51a42d94cd664233\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T00:07:09Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 00:07:03.713374 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 00:07:03.715052 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1528563048/tls.crt::/tmp/serving-cert-1528563048/tls.key\\\\\\\"\\\\nI1203 00:07:09.300052 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 00:07:09.304192 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 00:07:09.304255 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 00:07:09.304309 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 00:07:09.304323 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 00:07:09.313563 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 00:07:09.313597 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 00:07:09.313603 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 00:07:09.313609 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 00:07:09.313613 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 00:07:09.313618 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 00:07:09.313622 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 00:07:09.313709 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 00:07:09.327347 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T00:06:53Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c143d4232fb8bf7ef5999a664e431a00ab0db030f4d31a28d0bb21e00e38581\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:53Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72af6c27296eae2e275d455795c68f0d272869d63642b8353b17d466d0c901cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://72af6c27296eae2e275d455795c68f0d272869d63642b8353b17d466d0c901cf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:06:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:06:51Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:23Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:23 crc kubenswrapper[4953]: I1203 00:07:23.787359 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:10Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:23Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:23 crc kubenswrapper[4953]: I1203 00:07:23.800253 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce8b414e1fc22eb30decb6a6411b672e7f443055cb89ebb029fbf6ffc1d5dca7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:23Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:23 crc kubenswrapper[4953]: I1203 00:07:23.812454 4953 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 00:07:23 crc kubenswrapper[4953]: I1203 00:07:23.814056 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-kl5s9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d85e5c97-16ad-4bd7-9e73-b7f12678be97\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c21df91a9e4e3165686882b2eb559498224063dc633c3a63f958598108eca29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j24lb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:07:16Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-kl5s9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:23Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:23 crc kubenswrapper[4953]: I1203 00:07:23.815631 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:23 crc kubenswrapper[4953]: I1203 00:07:23.815670 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:23 crc kubenswrapper[4953]: I1203 00:07:23.815683 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:23 crc kubenswrapper[4953]: I1203 00:07:23.815698 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:23 crc kubenswrapper[4953]: I1203 00:07:23.815709 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:23Z","lastTransitionTime":"2025-12-03T00:07:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:23 crc kubenswrapper[4953]: I1203 00:07:23.834110 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f94f8c51-c665-4ca3-bf17-7a04ea5c7ade\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72c5cd7dc1c3a8adfedab6ba43a59d80b3efe65f2bb56be95e76cc86a409dcf3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2810ffc1b141fa7da811406053db9d72c708edf5f2bb1dd75e4d52ab35fd24c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c67b4277b20a9f1287f50ef82d0785c9d02543a8e7ed26e0413eae332b866db7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://258d78e1ad9d3dcda6976909ae3fe71aecead5b995fc66b3fd6e2722d80b5ab8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:06:51Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:23Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:23 crc kubenswrapper[4953]: I1203 00:07:23.851842 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09657a9f448a4c92a759c73a0a892076d7eb46393e6fd565e5f73f34c71aac23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5af5eac4a58870f63497f99b503cfa5fb922278095e0332ec50376925ecf2109\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:23Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:23 crc kubenswrapper[4953]: I1203 00:07:23.872622 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-wj6rz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"96aa67a0-15dd-420d-a1d5-ed239d7df1fb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01815cfd3f58c3f3f5f2a894c1790e0b69c5060c48dc0b1deff050b6ad7af916\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kf62m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b12bfb5bdf9fd4f2c50a335fc82df6ee0440eaf8741897da03aa2dae368cc6b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b12bfb5bdf9fd4f2c50a335fc82df6ee0440eaf8741897da03aa2dae368cc6b4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:07:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kf62m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83aaa5fa28110bc0ef05aff75cd1442bc9bbdd33b822ae2ef22d4e4ab83238bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://83aaa5fa28110bc0ef05aff75cd1442bc9bbdd33b822ae2ef22d4e4ab83238bc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:07:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:07:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kf62m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ab6213a33a19fedebce6d74a0ebbb3b3a2de5687504e21cc24093cc42ade635\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ab6213a33a19fedebce6d74a0ebbb3b3a2de5687504e21cc24093cc42ade635\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:07:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:07:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kf62m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://de480488843afa6b8dede1e1b42b5181dca1993e2f39dc1b461a18a57e3668e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://de480488843afa6b8dede1e1b42b5181dca1993e2f39dc1b461a18a57e3668e1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:07:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kf62m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://97be4ba692bc61078d4b0ba39fb0c12772548c6c9251ff3356ea702f151223cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://97be4ba692bc61078d4b0ba39fb0c12772548c6c9251ff3356ea702f151223cd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:07:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kf62m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1a1f009fcbd8af9217db1dc97f662719f8c1db232df0e2804e44c881abb1df2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b1a1f009fcbd8af9217db1dc97f662719f8c1db232df0e2804e44c881abb1df2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:07:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:07:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kf62m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:07:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-wj6rz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:23Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:23 crc kubenswrapper[4953]: I1203 00:07:23.902699 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-txffx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8816ebee-c735-4b8d-b1ec-65349707b347\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:16Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:16Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4b85224419a4075acb340e4b11095f2adb5c316e285afa4e19295c2181c0d24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0f6f645a58993f052a6f365bb808f62d99f7fea74886c5cc022b2d86112321a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a61b951a9cbab2c48c0f2df8351f97d7ff71bf3674f641fe5eac30155232b224\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd5f428de229749485267e81d34e31b708728b52474032e88460e54747f21e11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://00a2ea4a9bb870fce30e7fa845f787f184e7a19e42c95511655ea8c5f84a26d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc0a5d4cc9d614d8492626a1b114bdc52f510eb5965d842c02c083f08d4849a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aecc0ae6dd34893c9023da3b94ca01ae96ac1e7cd7f5fd18748cce6c273a7f9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7deadc8e135ff1bc7c5a2f093e489bb5b4b6f166bc874c6acc2a841c5397f11e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af4c7764fca87144d01f3738ea010308744ccf50c54b7c471ccc213112de00d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af4c7764fca87144d01f3738ea010308744ccf50c54b7c471ccc213112de00d5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:07:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:07:16Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-txffx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:23Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:23 crc kubenswrapper[4953]: I1203 00:07:23.917050 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5cmlc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7e11a1d3-ecee-4bf0-87bf-085fab97b73d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d42390b9c53efc63add8ddcebf03c976f5e5448eaa2c34e076b2dc90dbac60fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6x2sc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:07:18Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5cmlc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:23Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:23 crc kubenswrapper[4953]: I1203 00:07:23.918371 4953 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-txffx" Dec 03 00:07:23 crc kubenswrapper[4953]: I1203 00:07:23.922457 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:23 crc kubenswrapper[4953]: I1203 00:07:23.922485 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:23 crc kubenswrapper[4953]: I1203 00:07:23.922495 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:23 crc kubenswrapper[4953]: I1203 00:07:23.922511 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:23 crc kubenswrapper[4953]: I1203 00:07:23.922524 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:23Z","lastTransitionTime":"2025-12-03T00:07:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:23 crc kubenswrapper[4953]: I1203 00:07:23.935418 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cf821840-5353-4620-8f73-99b56b371ab0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cae5e2994fcc24dc9b2ea73e79572615b3311edc42642b7f278fb84123989b23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://29e4b9d1b70a93fb133e0759ad148ae133766a66efeae3ff06a21dfae0e805ff\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dec43536772aa7c3f5a662fa7f3bd73b1cae2791ad454f85792287828d0ffeed\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://034e0bac6892d361700b8f33d6427b8a55e2bd3f89b14ea77a374c6be94d25e2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://761973393a93c59c23e69f41b8ac0d16abc35abd61ecbc6d51a42d94cd664233\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T00:07:09Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 00:07:03.713374 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 00:07:03.715052 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1528563048/tls.crt::/tmp/serving-cert-1528563048/tls.key\\\\\\\"\\\\nI1203 00:07:09.300052 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 00:07:09.304192 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 00:07:09.304255 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 00:07:09.304309 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 00:07:09.304323 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 00:07:09.313563 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 00:07:09.313597 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 00:07:09.313603 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 00:07:09.313609 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 00:07:09.313613 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 00:07:09.313618 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 00:07:09.313622 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 00:07:09.313709 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 00:07:09.327347 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T00:06:53Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c143d4232fb8bf7ef5999a664e431a00ab0db030f4d31a28d0bb21e00e38581\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:53Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72af6c27296eae2e275d455795c68f0d272869d63642b8353b17d466d0c901cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://72af6c27296eae2e275d455795c68f0d272869d63642b8353b17d466d0c901cf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:06:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:06:51Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:23Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:23 crc kubenswrapper[4953]: I1203 00:07:23.949465 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:10Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:23Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:23 crc kubenswrapper[4953]: I1203 00:07:23.963932 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce8b414e1fc22eb30decb6a6411b672e7f443055cb89ebb029fbf6ffc1d5dca7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:23Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:23 crc kubenswrapper[4953]: I1203 00:07:23.979500 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-6dfbn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53ffb1e9-9a2e-4869-80d7-776f4aa7ef8b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ab4205c0d38ad6a0c32077eadb1fd5fd3a279d9c8b8a2ffbe693ec2c4018370c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ct26c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7df84405eda520ad5a22d35750edf07ef3c3977a756ad5739c54f91cba9da7b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ct26c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:07:16Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-6dfbn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:23Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:23 crc kubenswrapper[4953]: I1203 00:07:23.993725 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:23Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:24 crc kubenswrapper[4953]: I1203 00:07:24.008406 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2ed9c7fa96c8951fbc1728fe7a91fefdd999077285621df9c1abedf9ae0c449\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:24Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:24 crc kubenswrapper[4953]: I1203 00:07:24.021114 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:10Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:24Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:24 crc kubenswrapper[4953]: I1203 00:07:24.024985 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:24 crc kubenswrapper[4953]: I1203 00:07:24.025045 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:24 crc kubenswrapper[4953]: I1203 00:07:24.025057 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:24 crc kubenswrapper[4953]: I1203 00:07:24.025080 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:24 crc kubenswrapper[4953]: I1203 00:07:24.025096 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:24Z","lastTransitionTime":"2025-12-03T00:07:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:24 crc kubenswrapper[4953]: I1203 00:07:24.038735 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-675qk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"43a4bad8-71f4-441a-8302-aeb526a986af\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5b5794fef6477aab5deed8f642e8cc50b37eb1fd0cc6f19c003f34ad31c1e3c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gdrw4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:07:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-675qk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:24Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:24 crc kubenswrapper[4953]: I1203 00:07:24.054780 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:24Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:24 crc kubenswrapper[4953]: I1203 00:07:24.067630 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2ed9c7fa96c8951fbc1728fe7a91fefdd999077285621df9c1abedf9ae0c449\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:24Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:24 crc kubenswrapper[4953]: I1203 00:07:24.086594 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:10Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:24Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:24 crc kubenswrapper[4953]: I1203 00:07:24.098065 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-675qk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"43a4bad8-71f4-441a-8302-aeb526a986af\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5b5794fef6477aab5deed8f642e8cc50b37eb1fd0cc6f19c003f34ad31c1e3c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gdrw4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:07:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-675qk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:24Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:24 crc kubenswrapper[4953]: I1203 00:07:24.115409 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f94f8c51-c665-4ca3-bf17-7a04ea5c7ade\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72c5cd7dc1c3a8adfedab6ba43a59d80b3efe65f2bb56be95e76cc86a409dcf3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2810ffc1b141fa7da811406053db9d72c708edf5f2bb1dd75e4d52ab35fd24c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c67b4277b20a9f1287f50ef82d0785c9d02543a8e7ed26e0413eae332b866db7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://258d78e1ad9d3dcda6976909ae3fe71aecead5b995fc66b3fd6e2722d80b5ab8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:06:51Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:24Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:24 crc kubenswrapper[4953]: I1203 00:07:24.128271 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:24 crc kubenswrapper[4953]: I1203 00:07:24.128417 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:24 crc kubenswrapper[4953]: I1203 00:07:24.128500 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:24 crc kubenswrapper[4953]: I1203 00:07:24.128598 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:24 crc kubenswrapper[4953]: I1203 00:07:24.128660 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:24Z","lastTransitionTime":"2025-12-03T00:07:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:24 crc kubenswrapper[4953]: I1203 00:07:24.131836 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09657a9f448a4c92a759c73a0a892076d7eb46393e6fd565e5f73f34c71aac23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5af5eac4a58870f63497f99b503cfa5fb922278095e0332ec50376925ecf2109\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:24Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:24 crc kubenswrapper[4953]: I1203 00:07:24.151686 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-wj6rz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"96aa67a0-15dd-420d-a1d5-ed239d7df1fb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01815cfd3f58c3f3f5f2a894c1790e0b69c5060c48dc0b1deff050b6ad7af916\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kf62m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b12bfb5bdf9fd4f2c50a335fc82df6ee0440eaf8741897da03aa2dae368cc6b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b12bfb5bdf9fd4f2c50a335fc82df6ee0440eaf8741897da03aa2dae368cc6b4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:07:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kf62m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83aaa5fa28110bc0ef05aff75cd1442bc9bbdd33b822ae2ef22d4e4ab83238bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://83aaa5fa28110bc0ef05aff75cd1442bc9bbdd33b822ae2ef22d4e4ab83238bc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:07:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:07:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kf62m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ab6213a33a19fedebce6d74a0ebbb3b3a2de5687504e21cc24093cc42ade635\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ab6213a33a19fedebce6d74a0ebbb3b3a2de5687504e21cc24093cc42ade635\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:07:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:07:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kf62m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://de480488843afa6b8dede1e1b42b5181dca1993e2f39dc1b461a18a57e3668e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://de480488843afa6b8dede1e1b42b5181dca1993e2f39dc1b461a18a57e3668e1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:07:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kf62m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://97be4ba692bc61078d4b0ba39fb0c12772548c6c9251ff3356ea702f151223cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://97be4ba692bc61078d4b0ba39fb0c12772548c6c9251ff3356ea702f151223cd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:07:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kf62m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1a1f009fcbd8af9217db1dc97f662719f8c1db232df0e2804e44c881abb1df2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b1a1f009fcbd8af9217db1dc97f662719f8c1db232df0e2804e44c881abb1df2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:07:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:07:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kf62m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:07:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-wj6rz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:24Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:24 crc kubenswrapper[4953]: I1203 00:07:24.164689 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-kl5s9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d85e5c97-16ad-4bd7-9e73-b7f12678be97\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c21df91a9e4e3165686882b2eb559498224063dc633c3a63f958598108eca29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j24lb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:07:16Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-kl5s9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:24Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:24 crc kubenswrapper[4953]: I1203 00:07:24.175071 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5cmlc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7e11a1d3-ecee-4bf0-87bf-085fab97b73d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d42390b9c53efc63add8ddcebf03c976f5e5448eaa2c34e076b2dc90dbac60fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6x2sc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:07:18Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5cmlc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:24Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:24 crc kubenswrapper[4953]: I1203 00:07:24.190511 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cf821840-5353-4620-8f73-99b56b371ab0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cae5e2994fcc24dc9b2ea73e79572615b3311edc42642b7f278fb84123989b23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://29e4b9d1b70a93fb133e0759ad148ae133766a66efeae3ff06a21dfae0e805ff\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dec43536772aa7c3f5a662fa7f3bd73b1cae2791ad454f85792287828d0ffeed\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://034e0bac6892d361700b8f33d6427b8a55e2bd3f89b14ea77a374c6be94d25e2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://761973393a93c59c23e69f41b8ac0d16abc35abd61ecbc6d51a42d94cd664233\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T00:07:09Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 00:07:03.713374 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 00:07:03.715052 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1528563048/tls.crt::/tmp/serving-cert-1528563048/tls.key\\\\\\\"\\\\nI1203 00:07:09.300052 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 00:07:09.304192 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 00:07:09.304255 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 00:07:09.304309 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 00:07:09.304323 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 00:07:09.313563 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 00:07:09.313597 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 00:07:09.313603 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 00:07:09.313609 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 00:07:09.313613 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 00:07:09.313618 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 00:07:09.313622 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 00:07:09.313709 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 00:07:09.327347 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T00:06:53Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c143d4232fb8bf7ef5999a664e431a00ab0db030f4d31a28d0bb21e00e38581\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:53Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72af6c27296eae2e275d455795c68f0d272869d63642b8353b17d466d0c901cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://72af6c27296eae2e275d455795c68f0d272869d63642b8353b17d466d0c901cf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:06:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:06:51Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:24Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:24 crc kubenswrapper[4953]: I1203 00:07:24.205264 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:10Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:24Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:24 crc kubenswrapper[4953]: I1203 00:07:24.226122 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce8b414e1fc22eb30decb6a6411b672e7f443055cb89ebb029fbf6ffc1d5dca7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:24Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:24 crc kubenswrapper[4953]: I1203 00:07:24.231501 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:24 crc kubenswrapper[4953]: I1203 00:07:24.231543 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:24 crc kubenswrapper[4953]: I1203 00:07:24.231556 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:24 crc kubenswrapper[4953]: I1203 00:07:24.231577 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:24 crc kubenswrapper[4953]: I1203 00:07:24.231591 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:24Z","lastTransitionTime":"2025-12-03T00:07:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:24 crc kubenswrapper[4953]: I1203 00:07:24.239509 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-6dfbn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53ffb1e9-9a2e-4869-80d7-776f4aa7ef8b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ab4205c0d38ad6a0c32077eadb1fd5fd3a279d9c8b8a2ffbe693ec2c4018370c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ct26c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7df84405eda520ad5a22d35750edf07ef3c3977a756ad5739c54f91cba9da7b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ct26c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:07:16Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-6dfbn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:24Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:24 crc kubenswrapper[4953]: I1203 00:07:24.264746 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-txffx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8816ebee-c735-4b8d-b1ec-65349707b347\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:16Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:16Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4b85224419a4075acb340e4b11095f2adb5c316e285afa4e19295c2181c0d24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0f6f645a58993f052a6f365bb808f62d99f7fea74886c5cc022b2d86112321a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a61b951a9cbab2c48c0f2df8351f97d7ff71bf3674f641fe5eac30155232b224\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd5f428de229749485267e81d34e31b708728b52474032e88460e54747f21e11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://00a2ea4a9bb870fce30e7fa845f787f184e7a19e42c95511655ea8c5f84a26d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc0a5d4cc9d614d8492626a1b114bdc52f510eb5965d842c02c083f08d4849a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aecc0ae6dd34893c9023da3b94ca01ae96ac1e7cd7f5fd18748cce6c273a7f9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7deadc8e135ff1bc7c5a2f093e489bb5b4b6f166bc874c6acc2a841c5397f11e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af4c7764fca87144d01f3738ea010308744ccf50c54b7c471ccc213112de00d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af4c7764fca87144d01f3738ea010308744ccf50c54b7c471ccc213112de00d5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:07:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:07:16Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-txffx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:24Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:24 crc kubenswrapper[4953]: I1203 00:07:24.334208 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:24 crc kubenswrapper[4953]: I1203 00:07:24.334252 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:24 crc kubenswrapper[4953]: I1203 00:07:24.334265 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:24 crc kubenswrapper[4953]: I1203 00:07:24.334287 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:24 crc kubenswrapper[4953]: I1203 00:07:24.334303 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:24Z","lastTransitionTime":"2025-12-03T00:07:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:24 crc kubenswrapper[4953]: I1203 00:07:24.347054 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 00:07:24 crc kubenswrapper[4953]: E1203 00:07:24.347210 4953 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 00:07:24 crc kubenswrapper[4953]: I1203 00:07:24.347261 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 00:07:24 crc kubenswrapper[4953]: I1203 00:07:24.347374 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 00:07:24 crc kubenswrapper[4953]: E1203 00:07:24.347496 4953 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 00:07:24 crc kubenswrapper[4953]: E1203 00:07:24.347416 4953 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 00:07:24 crc kubenswrapper[4953]: I1203 00:07:24.437025 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:24 crc kubenswrapper[4953]: I1203 00:07:24.437089 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:24 crc kubenswrapper[4953]: I1203 00:07:24.437120 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:24 crc kubenswrapper[4953]: I1203 00:07:24.437147 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:24 crc kubenswrapper[4953]: I1203 00:07:24.437167 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:24Z","lastTransitionTime":"2025-12-03T00:07:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:24 crc kubenswrapper[4953]: I1203 00:07:24.540264 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:24 crc kubenswrapper[4953]: I1203 00:07:24.540325 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:24 crc kubenswrapper[4953]: I1203 00:07:24.540344 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:24 crc kubenswrapper[4953]: I1203 00:07:24.540371 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:24 crc kubenswrapper[4953]: I1203 00:07:24.540389 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:24Z","lastTransitionTime":"2025-12-03T00:07:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:24 crc kubenswrapper[4953]: I1203 00:07:24.628627 4953 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-txffx" Dec 03 00:07:24 crc kubenswrapper[4953]: I1203 00:07:24.629124 4953 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 03 00:07:24 crc kubenswrapper[4953]: I1203 00:07:24.643711 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:24 crc kubenswrapper[4953]: I1203 00:07:24.643751 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:24 crc kubenswrapper[4953]: I1203 00:07:24.643772 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:24 crc kubenswrapper[4953]: I1203 00:07:24.643795 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:24 crc kubenswrapper[4953]: I1203 00:07:24.643816 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:24Z","lastTransitionTime":"2025-12-03T00:07:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:24 crc kubenswrapper[4953]: I1203 00:07:24.710035 4953 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-txffx" Dec 03 00:07:24 crc kubenswrapper[4953]: I1203 00:07:24.730195 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:24Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:24 crc kubenswrapper[4953]: I1203 00:07:24.747268 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:24 crc kubenswrapper[4953]: I1203 00:07:24.747336 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:24 crc kubenswrapper[4953]: I1203 00:07:24.747357 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:24 crc kubenswrapper[4953]: I1203 00:07:24.747382 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:24 crc kubenswrapper[4953]: I1203 00:07:24.747402 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:24Z","lastTransitionTime":"2025-12-03T00:07:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:24 crc kubenswrapper[4953]: I1203 00:07:24.753676 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2ed9c7fa96c8951fbc1728fe7a91fefdd999077285621df9c1abedf9ae0c449\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:24Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:24 crc kubenswrapper[4953]: I1203 00:07:24.776603 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:10Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:24Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:24 crc kubenswrapper[4953]: I1203 00:07:24.798863 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-675qk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"43a4bad8-71f4-441a-8302-aeb526a986af\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5b5794fef6477aab5deed8f642e8cc50b37eb1fd0cc6f19c003f34ad31c1e3c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gdrw4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:07:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-675qk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:24Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:24 crc kubenswrapper[4953]: I1203 00:07:24.819771 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f94f8c51-c665-4ca3-bf17-7a04ea5c7ade\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72c5cd7dc1c3a8adfedab6ba43a59d80b3efe65f2bb56be95e76cc86a409dcf3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2810ffc1b141fa7da811406053db9d72c708edf5f2bb1dd75e4d52ab35fd24c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c67b4277b20a9f1287f50ef82d0785c9d02543a8e7ed26e0413eae332b866db7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://258d78e1ad9d3dcda6976909ae3fe71aecead5b995fc66b3fd6e2722d80b5ab8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:06:51Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:24Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:24 crc kubenswrapper[4953]: I1203 00:07:24.839477 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09657a9f448a4c92a759c73a0a892076d7eb46393e6fd565e5f73f34c71aac23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5af5eac4a58870f63497f99b503cfa5fb922278095e0332ec50376925ecf2109\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:24Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:24 crc kubenswrapper[4953]: I1203 00:07:24.850302 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:24 crc kubenswrapper[4953]: I1203 00:07:24.850440 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:24 crc kubenswrapper[4953]: I1203 00:07:24.850465 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:24 crc kubenswrapper[4953]: I1203 00:07:24.850495 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:24 crc kubenswrapper[4953]: I1203 00:07:24.850519 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:24Z","lastTransitionTime":"2025-12-03T00:07:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:24 crc kubenswrapper[4953]: I1203 00:07:24.858116 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-wj6rz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"96aa67a0-15dd-420d-a1d5-ed239d7df1fb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01815cfd3f58c3f3f5f2a894c1790e0b69c5060c48dc0b1deff050b6ad7af916\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kf62m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b12bfb5bdf9fd4f2c50a335fc82df6ee0440eaf8741897da03aa2dae368cc6b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b12bfb5bdf9fd4f2c50a335fc82df6ee0440eaf8741897da03aa2dae368cc6b4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:07:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kf62m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83aaa5fa28110bc0ef05aff75cd1442bc9bbdd33b822ae2ef22d4e4ab83238bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://83aaa5fa28110bc0ef05aff75cd1442bc9bbdd33b822ae2ef22d4e4ab83238bc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:07:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:07:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kf62m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ab6213a33a19fedebce6d74a0ebbb3b3a2de5687504e21cc24093cc42ade635\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ab6213a33a19fedebce6d74a0ebbb3b3a2de5687504e21cc24093cc42ade635\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:07:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:07:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kf62m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://de480488843afa6b8dede1e1b42b5181dca1993e2f39dc1b461a18a57e3668e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://de480488843afa6b8dede1e1b42b5181dca1993e2f39dc1b461a18a57e3668e1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:07:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kf62m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://97be4ba692bc61078d4b0ba39fb0c12772548c6c9251ff3356ea702f151223cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://97be4ba692bc61078d4b0ba39fb0c12772548c6c9251ff3356ea702f151223cd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:07:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kf62m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1a1f009fcbd8af9217db1dc97f662719f8c1db232df0e2804e44c881abb1df2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b1a1f009fcbd8af9217db1dc97f662719f8c1db232df0e2804e44c881abb1df2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:07:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:07:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kf62m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:07:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-wj6rz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:24Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:24 crc kubenswrapper[4953]: I1203 00:07:24.870929 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-kl5s9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d85e5c97-16ad-4bd7-9e73-b7f12678be97\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c21df91a9e4e3165686882b2eb559498224063dc633c3a63f958598108eca29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j24lb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:07:16Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-kl5s9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:24Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:24 crc kubenswrapper[4953]: I1203 00:07:24.907095 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cf821840-5353-4620-8f73-99b56b371ab0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cae5e2994fcc24dc9b2ea73e79572615b3311edc42642b7f278fb84123989b23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://29e4b9d1b70a93fb133e0759ad148ae133766a66efeae3ff06a21dfae0e805ff\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dec43536772aa7c3f5a662fa7f3bd73b1cae2791ad454f85792287828d0ffeed\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://034e0bac6892d361700b8f33d6427b8a55e2bd3f89b14ea77a374c6be94d25e2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://761973393a93c59c23e69f41b8ac0d16abc35abd61ecbc6d51a42d94cd664233\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T00:07:09Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 00:07:03.713374 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 00:07:03.715052 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1528563048/tls.crt::/tmp/serving-cert-1528563048/tls.key\\\\\\\"\\\\nI1203 00:07:09.300052 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 00:07:09.304192 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 00:07:09.304255 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 00:07:09.304309 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 00:07:09.304323 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 00:07:09.313563 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 00:07:09.313597 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 00:07:09.313603 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 00:07:09.313609 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 00:07:09.313613 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 00:07:09.313618 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 00:07:09.313622 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 00:07:09.313709 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 00:07:09.327347 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T00:06:53Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c143d4232fb8bf7ef5999a664e431a00ab0db030f4d31a28d0bb21e00e38581\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:53Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72af6c27296eae2e275d455795c68f0d272869d63642b8353b17d466d0c901cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://72af6c27296eae2e275d455795c68f0d272869d63642b8353b17d466d0c901cf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:06:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:06:51Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:24Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:24 crc kubenswrapper[4953]: I1203 00:07:24.951891 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:10Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:24Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:24 crc kubenswrapper[4953]: I1203 00:07:24.952580 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:24 crc kubenswrapper[4953]: I1203 00:07:24.952615 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:24 crc kubenswrapper[4953]: I1203 00:07:24.952626 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:24 crc kubenswrapper[4953]: I1203 00:07:24.952645 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:24 crc kubenswrapper[4953]: I1203 00:07:24.952657 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:24Z","lastTransitionTime":"2025-12-03T00:07:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:24 crc kubenswrapper[4953]: I1203 00:07:24.970455 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce8b414e1fc22eb30decb6a6411b672e7f443055cb89ebb029fbf6ffc1d5dca7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:24Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:24 crc kubenswrapper[4953]: I1203 00:07:24.980960 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-6dfbn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53ffb1e9-9a2e-4869-80d7-776f4aa7ef8b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ab4205c0d38ad6a0c32077eadb1fd5fd3a279d9c8b8a2ffbe693ec2c4018370c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ct26c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7df84405eda520ad5a22d35750edf07ef3c3977a756ad5739c54f91cba9da7b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ct26c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:07:16Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-6dfbn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:24Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:25 crc kubenswrapper[4953]: I1203 00:07:25.001361 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-txffx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8816ebee-c735-4b8d-b1ec-65349707b347\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:16Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:16Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4b85224419a4075acb340e4b11095f2adb5c316e285afa4e19295c2181c0d24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0f6f645a58993f052a6f365bb808f62d99f7fea74886c5cc022b2d86112321a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a61b951a9cbab2c48c0f2df8351f97d7ff71bf3674f641fe5eac30155232b224\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd5f428de229749485267e81d34e31b708728b52474032e88460e54747f21e11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://00a2ea4a9bb870fce30e7fa845f787f184e7a19e42c95511655ea8c5f84a26d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc0a5d4cc9d614d8492626a1b114bdc52f510eb5965d842c02c083f08d4849a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aecc0ae6dd34893c9023da3b94ca01ae96ac1e7cd7f5fd18748cce6c273a7f9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7deadc8e135ff1bc7c5a2f093e489bb5b4b6f166bc874c6acc2a841c5397f11e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af4c7764fca87144d01f3738ea010308744ccf50c54b7c471ccc213112de00d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af4c7764fca87144d01f3738ea010308744ccf50c54b7c471ccc213112de00d5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:07:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:07:16Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-txffx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:24Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:25 crc kubenswrapper[4953]: I1203 00:07:25.011174 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5cmlc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7e11a1d3-ecee-4bf0-87bf-085fab97b73d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d42390b9c53efc63add8ddcebf03c976f5e5448eaa2c34e076b2dc90dbac60fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6x2sc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:07:18Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5cmlc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:25Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:25 crc kubenswrapper[4953]: I1203 00:07:25.054822 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:25 crc kubenswrapper[4953]: I1203 00:07:25.054868 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:25 crc kubenswrapper[4953]: I1203 00:07:25.054878 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:25 crc kubenswrapper[4953]: I1203 00:07:25.054892 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:25 crc kubenswrapper[4953]: I1203 00:07:25.054902 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:25Z","lastTransitionTime":"2025-12-03T00:07:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:25 crc kubenswrapper[4953]: I1203 00:07:25.157867 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:25 crc kubenswrapper[4953]: I1203 00:07:25.158002 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:25 crc kubenswrapper[4953]: I1203 00:07:25.158042 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:25 crc kubenswrapper[4953]: I1203 00:07:25.158071 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:25 crc kubenswrapper[4953]: I1203 00:07:25.158094 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:25Z","lastTransitionTime":"2025-12-03T00:07:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:25 crc kubenswrapper[4953]: I1203 00:07:25.260395 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:25 crc kubenswrapper[4953]: I1203 00:07:25.261041 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:25 crc kubenswrapper[4953]: I1203 00:07:25.261156 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:25 crc kubenswrapper[4953]: I1203 00:07:25.261308 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:25 crc kubenswrapper[4953]: I1203 00:07:25.261428 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:25Z","lastTransitionTime":"2025-12-03T00:07:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:25 crc kubenswrapper[4953]: I1203 00:07:25.363948 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:25 crc kubenswrapper[4953]: I1203 00:07:25.364251 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:25 crc kubenswrapper[4953]: I1203 00:07:25.364353 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:25 crc kubenswrapper[4953]: I1203 00:07:25.364448 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:25 crc kubenswrapper[4953]: I1203 00:07:25.364525 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:25Z","lastTransitionTime":"2025-12-03T00:07:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:25 crc kubenswrapper[4953]: I1203 00:07:25.466639 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:25 crc kubenswrapper[4953]: I1203 00:07:25.466880 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:25 crc kubenswrapper[4953]: I1203 00:07:25.466955 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:25 crc kubenswrapper[4953]: I1203 00:07:25.467062 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:25 crc kubenswrapper[4953]: I1203 00:07:25.467158 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:25Z","lastTransitionTime":"2025-12-03T00:07:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:25 crc kubenswrapper[4953]: I1203 00:07:25.570719 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:25 crc kubenswrapper[4953]: I1203 00:07:25.570879 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:25 crc kubenswrapper[4953]: I1203 00:07:25.570959 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:25 crc kubenswrapper[4953]: I1203 00:07:25.571077 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:25 crc kubenswrapper[4953]: I1203 00:07:25.571181 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:25Z","lastTransitionTime":"2025-12-03T00:07:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:25 crc kubenswrapper[4953]: I1203 00:07:25.630761 4953 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 03 00:07:25 crc kubenswrapper[4953]: I1203 00:07:25.674232 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:25 crc kubenswrapper[4953]: I1203 00:07:25.674659 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:25 crc kubenswrapper[4953]: I1203 00:07:25.674836 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:25 crc kubenswrapper[4953]: I1203 00:07:25.675078 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:25 crc kubenswrapper[4953]: I1203 00:07:25.675301 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:25Z","lastTransitionTime":"2025-12-03T00:07:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:25 crc kubenswrapper[4953]: I1203 00:07:25.812066 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:25 crc kubenswrapper[4953]: I1203 00:07:25.812136 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:25 crc kubenswrapper[4953]: I1203 00:07:25.812158 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:25 crc kubenswrapper[4953]: I1203 00:07:25.812184 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:25 crc kubenswrapper[4953]: I1203 00:07:25.812202 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:25Z","lastTransitionTime":"2025-12-03T00:07:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:25 crc kubenswrapper[4953]: I1203 00:07:25.915440 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:25 crc kubenswrapper[4953]: I1203 00:07:25.915498 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:25 crc kubenswrapper[4953]: I1203 00:07:25.915511 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:25 crc kubenswrapper[4953]: I1203 00:07:25.915533 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:25 crc kubenswrapper[4953]: I1203 00:07:25.915545 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:25Z","lastTransitionTime":"2025-12-03T00:07:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:26 crc kubenswrapper[4953]: I1203 00:07:26.017659 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:26 crc kubenswrapper[4953]: I1203 00:07:26.017711 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:26 crc kubenswrapper[4953]: I1203 00:07:26.017725 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:26 crc kubenswrapper[4953]: I1203 00:07:26.017743 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:26 crc kubenswrapper[4953]: I1203 00:07:26.017754 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:26Z","lastTransitionTime":"2025-12-03T00:07:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:26 crc kubenswrapper[4953]: I1203 00:07:26.120795 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:26 crc kubenswrapper[4953]: I1203 00:07:26.120894 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:26 crc kubenswrapper[4953]: I1203 00:07:26.120912 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:26 crc kubenswrapper[4953]: I1203 00:07:26.120942 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:26 crc kubenswrapper[4953]: I1203 00:07:26.120962 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:26Z","lastTransitionTime":"2025-12-03T00:07:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:26 crc kubenswrapper[4953]: I1203 00:07:26.128343 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 00:07:26 crc kubenswrapper[4953]: I1203 00:07:26.128525 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 00:07:26 crc kubenswrapper[4953]: E1203 00:07:26.128570 4953 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 00:07:42.128536351 +0000 UTC m=+51.112817174 (durationBeforeRetry 16s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 00:07:26 crc kubenswrapper[4953]: I1203 00:07:26.128631 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 00:07:26 crc kubenswrapper[4953]: E1203 00:07:26.128713 4953 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 03 00:07:26 crc kubenswrapper[4953]: E1203 00:07:26.128742 4953 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 03 00:07:26 crc kubenswrapper[4953]: E1203 00:07:26.128763 4953 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 00:07:26 crc kubenswrapper[4953]: I1203 00:07:26.128715 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 00:07:26 crc kubenswrapper[4953]: E1203 00:07:26.128840 4953 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 03 00:07:26 crc kubenswrapper[4953]: E1203 00:07:26.128844 4953 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-03 00:07:42.128822288 +0000 UTC m=+51.113103101 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 00:07:26 crc kubenswrapper[4953]: E1203 00:07:26.128901 4953 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 03 00:07:26 crc kubenswrapper[4953]: I1203 00:07:26.129059 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 00:07:26 crc kubenswrapper[4953]: E1203 00:07:26.129075 4953 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 03 00:07:26 crc kubenswrapper[4953]: E1203 00:07:26.129143 4953 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 03 00:07:26 crc kubenswrapper[4953]: E1203 00:07:26.129145 4953 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 00:07:26 crc kubenswrapper[4953]: E1203 00:07:26.129159 4953 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-03 00:07:42.129113565 +0000 UTC m=+51.113394418 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 03 00:07:26 crc kubenswrapper[4953]: E1203 00:07:26.129223 4953 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-03 00:07:42.129191747 +0000 UTC m=+51.113472560 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 03 00:07:26 crc kubenswrapper[4953]: E1203 00:07:26.129250 4953 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-03 00:07:42.129238098 +0000 UTC m=+51.113518911 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 00:07:26 crc kubenswrapper[4953]: I1203 00:07:26.224205 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:26 crc kubenswrapper[4953]: I1203 00:07:26.224267 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:26 crc kubenswrapper[4953]: I1203 00:07:26.224278 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:26 crc kubenswrapper[4953]: I1203 00:07:26.224298 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:26 crc kubenswrapper[4953]: I1203 00:07:26.224323 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:26Z","lastTransitionTime":"2025-12-03T00:07:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:26 crc kubenswrapper[4953]: I1203 00:07:26.327012 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:26 crc kubenswrapper[4953]: I1203 00:07:26.327093 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:26 crc kubenswrapper[4953]: I1203 00:07:26.327112 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:26 crc kubenswrapper[4953]: I1203 00:07:26.327149 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:26 crc kubenswrapper[4953]: I1203 00:07:26.327167 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:26Z","lastTransitionTime":"2025-12-03T00:07:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:26 crc kubenswrapper[4953]: I1203 00:07:26.347805 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 00:07:26 crc kubenswrapper[4953]: I1203 00:07:26.347905 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 00:07:26 crc kubenswrapper[4953]: I1203 00:07:26.347814 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 00:07:26 crc kubenswrapper[4953]: E1203 00:07:26.348076 4953 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 00:07:26 crc kubenswrapper[4953]: E1203 00:07:26.348267 4953 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 00:07:26 crc kubenswrapper[4953]: E1203 00:07:26.348448 4953 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 00:07:26 crc kubenswrapper[4953]: I1203 00:07:26.429784 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:26 crc kubenswrapper[4953]: I1203 00:07:26.429921 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:26 crc kubenswrapper[4953]: I1203 00:07:26.429952 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:26 crc kubenswrapper[4953]: I1203 00:07:26.430039 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:26 crc kubenswrapper[4953]: I1203 00:07:26.430069 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:26Z","lastTransitionTime":"2025-12-03T00:07:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:26 crc kubenswrapper[4953]: I1203 00:07:26.533297 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:26 crc kubenswrapper[4953]: I1203 00:07:26.533346 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:26 crc kubenswrapper[4953]: I1203 00:07:26.533362 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:26 crc kubenswrapper[4953]: I1203 00:07:26.533387 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:26 crc kubenswrapper[4953]: I1203 00:07:26.533405 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:26Z","lastTransitionTime":"2025-12-03T00:07:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:26 crc kubenswrapper[4953]: I1203 00:07:26.633545 4953 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 03 00:07:26 crc kubenswrapper[4953]: I1203 00:07:26.635773 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:26 crc kubenswrapper[4953]: I1203 00:07:26.635844 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:26 crc kubenswrapper[4953]: I1203 00:07:26.635860 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:26 crc kubenswrapper[4953]: I1203 00:07:26.635880 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:26 crc kubenswrapper[4953]: I1203 00:07:26.635893 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:26Z","lastTransitionTime":"2025-12-03T00:07:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:26 crc kubenswrapper[4953]: I1203 00:07:26.738209 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:26 crc kubenswrapper[4953]: I1203 00:07:26.738284 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:26 crc kubenswrapper[4953]: I1203 00:07:26.738307 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:26 crc kubenswrapper[4953]: I1203 00:07:26.738343 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:26 crc kubenswrapper[4953]: I1203 00:07:26.738365 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:26Z","lastTransitionTime":"2025-12-03T00:07:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:26 crc kubenswrapper[4953]: I1203 00:07:26.841470 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:26 crc kubenswrapper[4953]: I1203 00:07:26.841567 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:26 crc kubenswrapper[4953]: I1203 00:07:26.841590 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:26 crc kubenswrapper[4953]: I1203 00:07:26.841670 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:26 crc kubenswrapper[4953]: I1203 00:07:26.841695 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:26Z","lastTransitionTime":"2025-12-03T00:07:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:26 crc kubenswrapper[4953]: I1203 00:07:26.945457 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:26 crc kubenswrapper[4953]: I1203 00:07:26.945513 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:26 crc kubenswrapper[4953]: I1203 00:07:26.945524 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:26 crc kubenswrapper[4953]: I1203 00:07:26.945544 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:26 crc kubenswrapper[4953]: I1203 00:07:26.945558 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:26Z","lastTransitionTime":"2025-12-03T00:07:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:27 crc kubenswrapper[4953]: I1203 00:07:27.048462 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:27 crc kubenswrapper[4953]: I1203 00:07:27.048522 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:27 crc kubenswrapper[4953]: I1203 00:07:27.048539 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:27 crc kubenswrapper[4953]: I1203 00:07:27.048563 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:27 crc kubenswrapper[4953]: I1203 00:07:27.048580 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:27Z","lastTransitionTime":"2025-12-03T00:07:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:27 crc kubenswrapper[4953]: I1203 00:07:27.152130 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:27 crc kubenswrapper[4953]: I1203 00:07:27.152193 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:27 crc kubenswrapper[4953]: I1203 00:07:27.152214 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:27 crc kubenswrapper[4953]: I1203 00:07:27.152248 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:27 crc kubenswrapper[4953]: I1203 00:07:27.152267 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:27Z","lastTransitionTime":"2025-12-03T00:07:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:27 crc kubenswrapper[4953]: I1203 00:07:27.255670 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:27 crc kubenswrapper[4953]: I1203 00:07:27.255728 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:27 crc kubenswrapper[4953]: I1203 00:07:27.255746 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:27 crc kubenswrapper[4953]: I1203 00:07:27.255769 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:27 crc kubenswrapper[4953]: I1203 00:07:27.255822 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:27Z","lastTransitionTime":"2025-12-03T00:07:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:27 crc kubenswrapper[4953]: I1203 00:07:27.359433 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:27 crc kubenswrapper[4953]: I1203 00:07:27.359509 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:27 crc kubenswrapper[4953]: I1203 00:07:27.359533 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:27 crc kubenswrapper[4953]: I1203 00:07:27.359562 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:27 crc kubenswrapper[4953]: I1203 00:07:27.359584 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:27Z","lastTransitionTime":"2025-12-03T00:07:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:27 crc kubenswrapper[4953]: I1203 00:07:27.462889 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:27 crc kubenswrapper[4953]: I1203 00:07:27.462953 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:27 crc kubenswrapper[4953]: I1203 00:07:27.462970 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:27 crc kubenswrapper[4953]: I1203 00:07:27.463020 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:27 crc kubenswrapper[4953]: I1203 00:07:27.463037 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:27Z","lastTransitionTime":"2025-12-03T00:07:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:27 crc kubenswrapper[4953]: I1203 00:07:27.566451 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:27 crc kubenswrapper[4953]: I1203 00:07:27.566518 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:27 crc kubenswrapper[4953]: I1203 00:07:27.566537 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:27 crc kubenswrapper[4953]: I1203 00:07:27.566560 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:27 crc kubenswrapper[4953]: I1203 00:07:27.566580 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:27Z","lastTransitionTime":"2025-12-03T00:07:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:27 crc kubenswrapper[4953]: I1203 00:07:27.640485 4953 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-txffx_8816ebee-c735-4b8d-b1ec-65349707b347/ovnkube-controller/0.log" Dec 03 00:07:27 crc kubenswrapper[4953]: I1203 00:07:27.645800 4953 generic.go:334] "Generic (PLEG): container finished" podID="8816ebee-c735-4b8d-b1ec-65349707b347" containerID="aecc0ae6dd34893c9023da3b94ca01ae96ac1e7cd7f5fd18748cce6c273a7f9e" exitCode=1 Dec 03 00:07:27 crc kubenswrapper[4953]: I1203 00:07:27.645853 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-txffx" event={"ID":"8816ebee-c735-4b8d-b1ec-65349707b347","Type":"ContainerDied","Data":"aecc0ae6dd34893c9023da3b94ca01ae96ac1e7cd7f5fd18748cce6c273a7f9e"} Dec 03 00:07:27 crc kubenswrapper[4953]: I1203 00:07:27.646949 4953 scope.go:117] "RemoveContainer" containerID="aecc0ae6dd34893c9023da3b94ca01ae96ac1e7cd7f5fd18748cce6c273a7f9e" Dec 03 00:07:27 crc kubenswrapper[4953]: I1203 00:07:27.669678 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:27 crc kubenswrapper[4953]: I1203 00:07:27.669734 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:27 crc kubenswrapper[4953]: I1203 00:07:27.669746 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:27 crc kubenswrapper[4953]: I1203 00:07:27.669763 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:27 crc kubenswrapper[4953]: I1203 00:07:27.669775 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:27Z","lastTransitionTime":"2025-12-03T00:07:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:27 crc kubenswrapper[4953]: I1203 00:07:27.676865 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:27Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:27 crc kubenswrapper[4953]: I1203 00:07:27.697727 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2ed9c7fa96c8951fbc1728fe7a91fefdd999077285621df9c1abedf9ae0c449\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:27Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:27 crc kubenswrapper[4953]: I1203 00:07:27.717263 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:10Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:27Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:27 crc kubenswrapper[4953]: I1203 00:07:27.734318 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-675qk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"43a4bad8-71f4-441a-8302-aeb526a986af\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5b5794fef6477aab5deed8f642e8cc50b37eb1fd0cc6f19c003f34ad31c1e3c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gdrw4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:07:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-675qk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:27Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:27 crc kubenswrapper[4953]: I1203 00:07:27.753735 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f94f8c51-c665-4ca3-bf17-7a04ea5c7ade\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72c5cd7dc1c3a8adfedab6ba43a59d80b3efe65f2bb56be95e76cc86a409dcf3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2810ffc1b141fa7da811406053db9d72c708edf5f2bb1dd75e4d52ab35fd24c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c67b4277b20a9f1287f50ef82d0785c9d02543a8e7ed26e0413eae332b866db7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://258d78e1ad9d3dcda6976909ae3fe71aecead5b995fc66b3fd6e2722d80b5ab8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:06:51Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:27Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:27 crc kubenswrapper[4953]: I1203 00:07:27.770601 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09657a9f448a4c92a759c73a0a892076d7eb46393e6fd565e5f73f34c71aac23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5af5eac4a58870f63497f99b503cfa5fb922278095e0332ec50376925ecf2109\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:27Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:27 crc kubenswrapper[4953]: I1203 00:07:27.772850 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:27 crc kubenswrapper[4953]: I1203 00:07:27.772958 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:27 crc kubenswrapper[4953]: I1203 00:07:27.772998 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:27 crc kubenswrapper[4953]: I1203 00:07:27.773020 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:27 crc kubenswrapper[4953]: I1203 00:07:27.773031 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:27Z","lastTransitionTime":"2025-12-03T00:07:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:27 crc kubenswrapper[4953]: I1203 00:07:27.789688 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-wj6rz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"96aa67a0-15dd-420d-a1d5-ed239d7df1fb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01815cfd3f58c3f3f5f2a894c1790e0b69c5060c48dc0b1deff050b6ad7af916\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kf62m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b12bfb5bdf9fd4f2c50a335fc82df6ee0440eaf8741897da03aa2dae368cc6b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b12bfb5bdf9fd4f2c50a335fc82df6ee0440eaf8741897da03aa2dae368cc6b4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:07:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kf62m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83aaa5fa28110bc0ef05aff75cd1442bc9bbdd33b822ae2ef22d4e4ab83238bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://83aaa5fa28110bc0ef05aff75cd1442bc9bbdd33b822ae2ef22d4e4ab83238bc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:07:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:07:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kf62m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ab6213a33a19fedebce6d74a0ebbb3b3a2de5687504e21cc24093cc42ade635\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ab6213a33a19fedebce6d74a0ebbb3b3a2de5687504e21cc24093cc42ade635\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:07:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:07:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kf62m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://de480488843afa6b8dede1e1b42b5181dca1993e2f39dc1b461a18a57e3668e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://de480488843afa6b8dede1e1b42b5181dca1993e2f39dc1b461a18a57e3668e1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:07:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kf62m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://97be4ba692bc61078d4b0ba39fb0c12772548c6c9251ff3356ea702f151223cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://97be4ba692bc61078d4b0ba39fb0c12772548c6c9251ff3356ea702f151223cd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:07:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kf62m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1a1f009fcbd8af9217db1dc97f662719f8c1db232df0e2804e44c881abb1df2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b1a1f009fcbd8af9217db1dc97f662719f8c1db232df0e2804e44c881abb1df2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:07:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:07:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kf62m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:07:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-wj6rz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:27Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:27 crc kubenswrapper[4953]: I1203 00:07:27.799888 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-kl5s9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d85e5c97-16ad-4bd7-9e73-b7f12678be97\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c21df91a9e4e3165686882b2eb559498224063dc633c3a63f958598108eca29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j24lb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:07:16Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-kl5s9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:27Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:27 crc kubenswrapper[4953]: I1203 00:07:27.812218 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cf821840-5353-4620-8f73-99b56b371ab0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cae5e2994fcc24dc9b2ea73e79572615b3311edc42642b7f278fb84123989b23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://29e4b9d1b70a93fb133e0759ad148ae133766a66efeae3ff06a21dfae0e805ff\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dec43536772aa7c3f5a662fa7f3bd73b1cae2791ad454f85792287828d0ffeed\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://034e0bac6892d361700b8f33d6427b8a55e2bd3f89b14ea77a374c6be94d25e2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://761973393a93c59c23e69f41b8ac0d16abc35abd61ecbc6d51a42d94cd664233\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T00:07:09Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 00:07:03.713374 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 00:07:03.715052 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1528563048/tls.crt::/tmp/serving-cert-1528563048/tls.key\\\\\\\"\\\\nI1203 00:07:09.300052 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 00:07:09.304192 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 00:07:09.304255 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 00:07:09.304309 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 00:07:09.304323 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 00:07:09.313563 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 00:07:09.313597 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 00:07:09.313603 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 00:07:09.313609 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 00:07:09.313613 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 00:07:09.313618 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 00:07:09.313622 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 00:07:09.313709 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 00:07:09.327347 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T00:06:53Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c143d4232fb8bf7ef5999a664e431a00ab0db030f4d31a28d0bb21e00e38581\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:53Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72af6c27296eae2e275d455795c68f0d272869d63642b8353b17d466d0c901cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://72af6c27296eae2e275d455795c68f0d272869d63642b8353b17d466d0c901cf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:06:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:06:51Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:27Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:27 crc kubenswrapper[4953]: I1203 00:07:27.822588 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:10Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:27Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:27 crc kubenswrapper[4953]: I1203 00:07:27.835382 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce8b414e1fc22eb30decb6a6411b672e7f443055cb89ebb029fbf6ffc1d5dca7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:27Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:27 crc kubenswrapper[4953]: I1203 00:07:27.845522 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-6dfbn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53ffb1e9-9a2e-4869-80d7-776f4aa7ef8b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ab4205c0d38ad6a0c32077eadb1fd5fd3a279d9c8b8a2ffbe693ec2c4018370c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ct26c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7df84405eda520ad5a22d35750edf07ef3c3977a756ad5739c54f91cba9da7b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ct26c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:07:16Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-6dfbn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:27Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:27 crc kubenswrapper[4953]: I1203 00:07:27.862340 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-txffx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8816ebee-c735-4b8d-b1ec-65349707b347\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:16Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:16Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4b85224419a4075acb340e4b11095f2adb5c316e285afa4e19295c2181c0d24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0f6f645a58993f052a6f365bb808f62d99f7fea74886c5cc022b2d86112321a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a61b951a9cbab2c48c0f2df8351f97d7ff71bf3674f641fe5eac30155232b224\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd5f428de229749485267e81d34e31b708728b52474032e88460e54747f21e11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://00a2ea4a9bb870fce30e7fa845f787f184e7a19e42c95511655ea8c5f84a26d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc0a5d4cc9d614d8492626a1b114bdc52f510eb5965d842c02c083f08d4849a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aecc0ae6dd34893c9023da3b94ca01ae96ac1e7cd7f5fd18748cce6c273a7f9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aecc0ae6dd34893c9023da3b94ca01ae96ac1e7cd7f5fd18748cce6c273a7f9e\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T00:07:26Z\\\",\\\"message\\\":\\\"ller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1203 00:07:25.992760 6257 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1203 00:07:25.992789 6257 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1203 00:07:25.992837 6257 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1203 00:07:25.992889 6257 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1203 00:07:25.992927 6257 handler.go:208] Removed *v1.Node event handler 7\\\\nI1203 00:07:25.992967 6257 handler.go:208] Removed *v1.Node event handler 2\\\\nI1203 00:07:25.992993 6257 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1203 00:07:25.992947 6257 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1203 00:07:25.993079 6257 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1203 00:07:25.992969 6257 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1203 00:07:25.993139 6257 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1203 00:07:25.993168 6257 factory.go:656] Stopping watch factory\\\\nI1203 00:07:25.993190 6257 ovnkube.go:599] Stopped ovnkube\\\\nI1203 00:07:25.993215 6257 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1203 00:07:25.993246 6257 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T00:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7deadc8e135ff1bc7c5a2f093e489bb5b4b6f166bc874c6acc2a841c5397f11e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af4c7764fca87144d01f3738ea010308744ccf50c54b7c471ccc213112de00d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af4c7764fca87144d01f3738ea010308744ccf50c54b7c471ccc213112de00d5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:07:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:07:16Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-txffx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:27Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:27 crc kubenswrapper[4953]: I1203 00:07:27.872542 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5cmlc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7e11a1d3-ecee-4bf0-87bf-085fab97b73d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d42390b9c53efc63add8ddcebf03c976f5e5448eaa2c34e076b2dc90dbac60fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6x2sc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:07:18Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5cmlc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:27Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:27 crc kubenswrapper[4953]: I1203 00:07:27.875376 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:27 crc kubenswrapper[4953]: I1203 00:07:27.875442 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:27 crc kubenswrapper[4953]: I1203 00:07:27.875465 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:27 crc kubenswrapper[4953]: I1203 00:07:27.875494 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:27 crc kubenswrapper[4953]: I1203 00:07:27.875518 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:27Z","lastTransitionTime":"2025-12-03T00:07:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:27 crc kubenswrapper[4953]: I1203 00:07:27.978628 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:27 crc kubenswrapper[4953]: I1203 00:07:27.978669 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:27 crc kubenswrapper[4953]: I1203 00:07:27.978679 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:27 crc kubenswrapper[4953]: I1203 00:07:27.978696 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:27 crc kubenswrapper[4953]: I1203 00:07:27.978718 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:27Z","lastTransitionTime":"2025-12-03T00:07:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:28 crc kubenswrapper[4953]: I1203 00:07:28.080878 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:28 crc kubenswrapper[4953]: I1203 00:07:28.080908 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:28 crc kubenswrapper[4953]: I1203 00:07:28.080915 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:28 crc kubenswrapper[4953]: I1203 00:07:28.080928 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:28 crc kubenswrapper[4953]: I1203 00:07:28.080936 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:28Z","lastTransitionTime":"2025-12-03T00:07:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:28 crc kubenswrapper[4953]: I1203 00:07:28.183466 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:28 crc kubenswrapper[4953]: I1203 00:07:28.183607 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:28 crc kubenswrapper[4953]: I1203 00:07:28.183633 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:28 crc kubenswrapper[4953]: I1203 00:07:28.183661 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:28 crc kubenswrapper[4953]: I1203 00:07:28.183685 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:28Z","lastTransitionTime":"2025-12-03T00:07:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:28 crc kubenswrapper[4953]: I1203 00:07:28.286854 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:28 crc kubenswrapper[4953]: I1203 00:07:28.286928 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:28 crc kubenswrapper[4953]: I1203 00:07:28.286955 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:28 crc kubenswrapper[4953]: I1203 00:07:28.287022 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:28 crc kubenswrapper[4953]: I1203 00:07:28.287050 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:28Z","lastTransitionTime":"2025-12-03T00:07:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:28 crc kubenswrapper[4953]: I1203 00:07:28.347539 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 00:07:28 crc kubenswrapper[4953]: I1203 00:07:28.347579 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 00:07:28 crc kubenswrapper[4953]: E1203 00:07:28.347646 4953 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 00:07:28 crc kubenswrapper[4953]: I1203 00:07:28.347672 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 00:07:28 crc kubenswrapper[4953]: E1203 00:07:28.347836 4953 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 00:07:28 crc kubenswrapper[4953]: E1203 00:07:28.347889 4953 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 00:07:28 crc kubenswrapper[4953]: I1203 00:07:28.391741 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:28 crc kubenswrapper[4953]: I1203 00:07:28.391830 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:28 crc kubenswrapper[4953]: I1203 00:07:28.391866 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:28 crc kubenswrapper[4953]: I1203 00:07:28.391897 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:28 crc kubenswrapper[4953]: I1203 00:07:28.391919 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:28Z","lastTransitionTime":"2025-12-03T00:07:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:28 crc kubenswrapper[4953]: I1203 00:07:28.500321 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:28 crc kubenswrapper[4953]: I1203 00:07:28.500374 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:28 crc kubenswrapper[4953]: I1203 00:07:28.500392 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:28 crc kubenswrapper[4953]: I1203 00:07:28.500416 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:28 crc kubenswrapper[4953]: I1203 00:07:28.500434 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:28Z","lastTransitionTime":"2025-12-03T00:07:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:28 crc kubenswrapper[4953]: I1203 00:07:28.603882 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:28 crc kubenswrapper[4953]: I1203 00:07:28.603932 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:28 crc kubenswrapper[4953]: I1203 00:07:28.603949 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:28 crc kubenswrapper[4953]: I1203 00:07:28.604025 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:28 crc kubenswrapper[4953]: I1203 00:07:28.604052 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:28Z","lastTransitionTime":"2025-12-03T00:07:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:28 crc kubenswrapper[4953]: I1203 00:07:28.652536 4953 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-txffx_8816ebee-c735-4b8d-b1ec-65349707b347/ovnkube-controller/0.log" Dec 03 00:07:28 crc kubenswrapper[4953]: I1203 00:07:28.657863 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-txffx" event={"ID":"8816ebee-c735-4b8d-b1ec-65349707b347","Type":"ContainerStarted","Data":"efbfc74a8598c47a658c06b597fa36d2a5ac8d02265d2864fd2832dd11ecb2cc"} Dec 03 00:07:28 crc kubenswrapper[4953]: I1203 00:07:28.708615 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:28 crc kubenswrapper[4953]: I1203 00:07:28.708672 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:28 crc kubenswrapper[4953]: I1203 00:07:28.708691 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:28 crc kubenswrapper[4953]: I1203 00:07:28.708721 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:28 crc kubenswrapper[4953]: I1203 00:07:28.708743 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:28Z","lastTransitionTime":"2025-12-03T00:07:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:28 crc kubenswrapper[4953]: I1203 00:07:28.811380 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:28 crc kubenswrapper[4953]: I1203 00:07:28.811417 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:28 crc kubenswrapper[4953]: I1203 00:07:28.811428 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:28 crc kubenswrapper[4953]: I1203 00:07:28.811444 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:28 crc kubenswrapper[4953]: I1203 00:07:28.811456 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:28Z","lastTransitionTime":"2025-12-03T00:07:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:28 crc kubenswrapper[4953]: I1203 00:07:28.913674 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:28 crc kubenswrapper[4953]: I1203 00:07:28.913716 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:28 crc kubenswrapper[4953]: I1203 00:07:28.913727 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:28 crc kubenswrapper[4953]: I1203 00:07:28.913744 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:28 crc kubenswrapper[4953]: I1203 00:07:28.913756 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:28Z","lastTransitionTime":"2025-12-03T00:07:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:29 crc kubenswrapper[4953]: I1203 00:07:29.016383 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:29 crc kubenswrapper[4953]: I1203 00:07:29.016451 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:29 crc kubenswrapper[4953]: I1203 00:07:29.016470 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:29 crc kubenswrapper[4953]: I1203 00:07:29.016496 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:29 crc kubenswrapper[4953]: I1203 00:07:29.016518 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:29Z","lastTransitionTime":"2025-12-03T00:07:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:29 crc kubenswrapper[4953]: I1203 00:07:29.119383 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:29 crc kubenswrapper[4953]: I1203 00:07:29.119440 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:29 crc kubenswrapper[4953]: I1203 00:07:29.119453 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:29 crc kubenswrapper[4953]: I1203 00:07:29.119474 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:29 crc kubenswrapper[4953]: I1203 00:07:29.119485 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:29Z","lastTransitionTime":"2025-12-03T00:07:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:29 crc kubenswrapper[4953]: I1203 00:07:29.222675 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:29 crc kubenswrapper[4953]: I1203 00:07:29.222731 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:29 crc kubenswrapper[4953]: I1203 00:07:29.222752 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:29 crc kubenswrapper[4953]: I1203 00:07:29.222783 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:29 crc kubenswrapper[4953]: I1203 00:07:29.222809 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:29Z","lastTransitionTime":"2025-12-03T00:07:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:29 crc kubenswrapper[4953]: I1203 00:07:29.325256 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:29 crc kubenswrapper[4953]: I1203 00:07:29.325316 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:29 crc kubenswrapper[4953]: I1203 00:07:29.325333 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:29 crc kubenswrapper[4953]: I1203 00:07:29.325357 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:29 crc kubenswrapper[4953]: I1203 00:07:29.325373 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:29Z","lastTransitionTime":"2025-12-03T00:07:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:29 crc kubenswrapper[4953]: I1203 00:07:29.428248 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:29 crc kubenswrapper[4953]: I1203 00:07:29.428323 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:29 crc kubenswrapper[4953]: I1203 00:07:29.428334 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:29 crc kubenswrapper[4953]: I1203 00:07:29.428351 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:29 crc kubenswrapper[4953]: I1203 00:07:29.428365 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:29Z","lastTransitionTime":"2025-12-03T00:07:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:29 crc kubenswrapper[4953]: I1203 00:07:29.530924 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:29 crc kubenswrapper[4953]: I1203 00:07:29.530988 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:29 crc kubenswrapper[4953]: I1203 00:07:29.531003 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:29 crc kubenswrapper[4953]: I1203 00:07:29.531027 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:29 crc kubenswrapper[4953]: I1203 00:07:29.531064 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:29Z","lastTransitionTime":"2025-12-03T00:07:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:29 crc kubenswrapper[4953]: I1203 00:07:29.633887 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:29 crc kubenswrapper[4953]: I1203 00:07:29.633928 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:29 crc kubenswrapper[4953]: I1203 00:07:29.633941 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:29 crc kubenswrapper[4953]: I1203 00:07:29.633959 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:29 crc kubenswrapper[4953]: I1203 00:07:29.633987 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:29Z","lastTransitionTime":"2025-12-03T00:07:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:29 crc kubenswrapper[4953]: I1203 00:07:29.660731 4953 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 03 00:07:29 crc kubenswrapper[4953]: I1203 00:07:29.678364 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cf821840-5353-4620-8f73-99b56b371ab0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cae5e2994fcc24dc9b2ea73e79572615b3311edc42642b7f278fb84123989b23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://29e4b9d1b70a93fb133e0759ad148ae133766a66efeae3ff06a21dfae0e805ff\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dec43536772aa7c3f5a662fa7f3bd73b1cae2791ad454f85792287828d0ffeed\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://034e0bac6892d361700b8f33d6427b8a55e2bd3f89b14ea77a374c6be94d25e2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://761973393a93c59c23e69f41b8ac0d16abc35abd61ecbc6d51a42d94cd664233\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T00:07:09Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 00:07:03.713374 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 00:07:03.715052 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1528563048/tls.crt::/tmp/serving-cert-1528563048/tls.key\\\\\\\"\\\\nI1203 00:07:09.300052 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 00:07:09.304192 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 00:07:09.304255 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 00:07:09.304309 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 00:07:09.304323 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 00:07:09.313563 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 00:07:09.313597 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 00:07:09.313603 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 00:07:09.313609 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 00:07:09.313613 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 00:07:09.313618 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 00:07:09.313622 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 00:07:09.313709 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 00:07:09.327347 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T00:06:53Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c143d4232fb8bf7ef5999a664e431a00ab0db030f4d31a28d0bb21e00e38581\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:53Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72af6c27296eae2e275d455795c68f0d272869d63642b8353b17d466d0c901cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://72af6c27296eae2e275d455795c68f0d272869d63642b8353b17d466d0c901cf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:06:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:06:51Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:29Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:29 crc kubenswrapper[4953]: I1203 00:07:29.692292 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:10Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:29Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:29 crc kubenswrapper[4953]: I1203 00:07:29.709124 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce8b414e1fc22eb30decb6a6411b672e7f443055cb89ebb029fbf6ffc1d5dca7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:29Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:29 crc kubenswrapper[4953]: I1203 00:07:29.721948 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-6dfbn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53ffb1e9-9a2e-4869-80d7-776f4aa7ef8b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ab4205c0d38ad6a0c32077eadb1fd5fd3a279d9c8b8a2ffbe693ec2c4018370c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ct26c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7df84405eda520ad5a22d35750edf07ef3c3977a756ad5739c54f91cba9da7b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ct26c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:07:16Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-6dfbn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:29Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:29 crc kubenswrapper[4953]: I1203 00:07:29.737042 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:29 crc kubenswrapper[4953]: I1203 00:07:29.737153 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:29 crc kubenswrapper[4953]: I1203 00:07:29.737176 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:29 crc kubenswrapper[4953]: I1203 00:07:29.737236 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:29 crc kubenswrapper[4953]: I1203 00:07:29.737256 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:29Z","lastTransitionTime":"2025-12-03T00:07:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:29 crc kubenswrapper[4953]: I1203 00:07:29.746765 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-txffx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8816ebee-c735-4b8d-b1ec-65349707b347\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:16Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:16Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4b85224419a4075acb340e4b11095f2adb5c316e285afa4e19295c2181c0d24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0f6f645a58993f052a6f365bb808f62d99f7fea74886c5cc022b2d86112321a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a61b951a9cbab2c48c0f2df8351f97d7ff71bf3674f641fe5eac30155232b224\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd5f428de229749485267e81d34e31b708728b52474032e88460e54747f21e11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://00a2ea4a9bb870fce30e7fa845f787f184e7a19e42c95511655ea8c5f84a26d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc0a5d4cc9d614d8492626a1b114bdc52f510eb5965d842c02c083f08d4849a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://efbfc74a8598c47a658c06b597fa36d2a5ac8d02265d2864fd2832dd11ecb2cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aecc0ae6dd34893c9023da3b94ca01ae96ac1e7cd7f5fd18748cce6c273a7f9e\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T00:07:26Z\\\",\\\"message\\\":\\\"ller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1203 00:07:25.992760 6257 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1203 00:07:25.992789 6257 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1203 00:07:25.992837 6257 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1203 00:07:25.992889 6257 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1203 00:07:25.992927 6257 handler.go:208] Removed *v1.Node event handler 7\\\\nI1203 00:07:25.992967 6257 handler.go:208] Removed *v1.Node event handler 2\\\\nI1203 00:07:25.992993 6257 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1203 00:07:25.992947 6257 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1203 00:07:25.993079 6257 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1203 00:07:25.992969 6257 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1203 00:07:25.993139 6257 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1203 00:07:25.993168 6257 factory.go:656] Stopping watch factory\\\\nI1203 00:07:25.993190 6257 ovnkube.go:599] Stopped ovnkube\\\\nI1203 00:07:25.993215 6257 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1203 00:07:25.993246 6257 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T00:07:23Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7deadc8e135ff1bc7c5a2f093e489bb5b4b6f166bc874c6acc2a841c5397f11e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af4c7764fca87144d01f3738ea010308744ccf50c54b7c471ccc213112de00d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af4c7764fca87144d01f3738ea010308744ccf50c54b7c471ccc213112de00d5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:07:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:07:16Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-txffx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:29Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:29 crc kubenswrapper[4953]: I1203 00:07:29.764194 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5cmlc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7e11a1d3-ecee-4bf0-87bf-085fab97b73d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d42390b9c53efc63add8ddcebf03c976f5e5448eaa2c34e076b2dc90dbac60fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6x2sc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:07:18Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5cmlc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:29Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:29 crc kubenswrapper[4953]: I1203 00:07:29.784020 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:29Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:29 crc kubenswrapper[4953]: I1203 00:07:29.801267 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2ed9c7fa96c8951fbc1728fe7a91fefdd999077285621df9c1abedf9ae0c449\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:29Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:29 crc kubenswrapper[4953]: I1203 00:07:29.817957 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:10Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:29Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:29 crc kubenswrapper[4953]: I1203 00:07:29.837591 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-675qk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"43a4bad8-71f4-441a-8302-aeb526a986af\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5b5794fef6477aab5deed8f642e8cc50b37eb1fd0cc6f19c003f34ad31c1e3c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gdrw4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:07:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-675qk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:29Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:29 crc kubenswrapper[4953]: I1203 00:07:29.840445 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:29 crc kubenswrapper[4953]: I1203 00:07:29.840496 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:29 crc kubenswrapper[4953]: I1203 00:07:29.840515 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:29 crc kubenswrapper[4953]: I1203 00:07:29.840541 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:29 crc kubenswrapper[4953]: I1203 00:07:29.840573 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:29Z","lastTransitionTime":"2025-12-03T00:07:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:29 crc kubenswrapper[4953]: I1203 00:07:29.853908 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f94f8c51-c665-4ca3-bf17-7a04ea5c7ade\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72c5cd7dc1c3a8adfedab6ba43a59d80b3efe65f2bb56be95e76cc86a409dcf3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2810ffc1b141fa7da811406053db9d72c708edf5f2bb1dd75e4d52ab35fd24c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c67b4277b20a9f1287f50ef82d0785c9d02543a8e7ed26e0413eae332b866db7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://258d78e1ad9d3dcda6976909ae3fe71aecead5b995fc66b3fd6e2722d80b5ab8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:06:51Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:29Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:29 crc kubenswrapper[4953]: I1203 00:07:29.876082 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09657a9f448a4c92a759c73a0a892076d7eb46393e6fd565e5f73f34c71aac23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5af5eac4a58870f63497f99b503cfa5fb922278095e0332ec50376925ecf2109\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:29Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:29 crc kubenswrapper[4953]: I1203 00:07:29.900409 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-wj6rz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"96aa67a0-15dd-420d-a1d5-ed239d7df1fb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01815cfd3f58c3f3f5f2a894c1790e0b69c5060c48dc0b1deff050b6ad7af916\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kf62m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b12bfb5bdf9fd4f2c50a335fc82df6ee0440eaf8741897da03aa2dae368cc6b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b12bfb5bdf9fd4f2c50a335fc82df6ee0440eaf8741897da03aa2dae368cc6b4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:07:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kf62m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83aaa5fa28110bc0ef05aff75cd1442bc9bbdd33b822ae2ef22d4e4ab83238bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://83aaa5fa28110bc0ef05aff75cd1442bc9bbdd33b822ae2ef22d4e4ab83238bc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:07:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:07:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kf62m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ab6213a33a19fedebce6d74a0ebbb3b3a2de5687504e21cc24093cc42ade635\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ab6213a33a19fedebce6d74a0ebbb3b3a2de5687504e21cc24093cc42ade635\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:07:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:07:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kf62m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://de480488843afa6b8dede1e1b42b5181dca1993e2f39dc1b461a18a57e3668e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://de480488843afa6b8dede1e1b42b5181dca1993e2f39dc1b461a18a57e3668e1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:07:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kf62m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://97be4ba692bc61078d4b0ba39fb0c12772548c6c9251ff3356ea702f151223cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://97be4ba692bc61078d4b0ba39fb0c12772548c6c9251ff3356ea702f151223cd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:07:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kf62m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1a1f009fcbd8af9217db1dc97f662719f8c1db232df0e2804e44c881abb1df2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b1a1f009fcbd8af9217db1dc97f662719f8c1db232df0e2804e44c881abb1df2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:07:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:07:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kf62m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:07:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-wj6rz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:29Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:29 crc kubenswrapper[4953]: I1203 00:07:29.914606 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-kl5s9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d85e5c97-16ad-4bd7-9e73-b7f12678be97\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c21df91a9e4e3165686882b2eb559498224063dc633c3a63f958598108eca29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j24lb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:07:16Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-kl5s9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:29Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:29 crc kubenswrapper[4953]: I1203 00:07:29.943732 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:29 crc kubenswrapper[4953]: I1203 00:07:29.943794 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:29 crc kubenswrapper[4953]: I1203 00:07:29.943817 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:29 crc kubenswrapper[4953]: I1203 00:07:29.943847 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:29 crc kubenswrapper[4953]: I1203 00:07:29.943871 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:29Z","lastTransitionTime":"2025-12-03T00:07:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:29 crc kubenswrapper[4953]: I1203 00:07:29.988289 4953 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-chmnz"] Dec 03 00:07:29 crc kubenswrapper[4953]: I1203 00:07:29.988939 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-chmnz" Dec 03 00:07:29 crc kubenswrapper[4953]: I1203 00:07:29.992763 4953 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Dec 03 00:07:29 crc kubenswrapper[4953]: I1203 00:07:29.993255 4953 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Dec 03 00:07:30 crc kubenswrapper[4953]: I1203 00:07:30.010089 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:30Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:30 crc kubenswrapper[4953]: I1203 00:07:30.029410 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2ed9c7fa96c8951fbc1728fe7a91fefdd999077285621df9c1abedf9ae0c449\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:30Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:30 crc kubenswrapper[4953]: I1203 00:07:30.048569 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:30 crc kubenswrapper[4953]: I1203 00:07:30.048688 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:30 crc kubenswrapper[4953]: I1203 00:07:30.048711 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:30 crc kubenswrapper[4953]: I1203 00:07:30.048740 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:30 crc kubenswrapper[4953]: I1203 00:07:30.048764 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:30Z","lastTransitionTime":"2025-12-03T00:07:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:30 crc kubenswrapper[4953]: I1203 00:07:30.048840 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:10Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:30Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:30 crc kubenswrapper[4953]: I1203 00:07:30.069063 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-675qk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"43a4bad8-71f4-441a-8302-aeb526a986af\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5b5794fef6477aab5deed8f642e8cc50b37eb1fd0cc6f19c003f34ad31c1e3c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gdrw4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:07:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-675qk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:30Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:30 crc kubenswrapper[4953]: I1203 00:07:30.073002 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/7a12fe6d-d25a-4b08-90da-cc9e510980a7-env-overrides\") pod \"ovnkube-control-plane-749d76644c-chmnz\" (UID: \"7a12fe6d-d25a-4b08-90da-cc9e510980a7\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-chmnz" Dec 03 00:07:30 crc kubenswrapper[4953]: I1203 00:07:30.073157 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/7a12fe6d-d25a-4b08-90da-cc9e510980a7-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-chmnz\" (UID: \"7a12fe6d-d25a-4b08-90da-cc9e510980a7\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-chmnz" Dec 03 00:07:30 crc kubenswrapper[4953]: I1203 00:07:30.073253 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c982g\" (UniqueName: \"kubernetes.io/projected/7a12fe6d-d25a-4b08-90da-cc9e510980a7-kube-api-access-c982g\") pod \"ovnkube-control-plane-749d76644c-chmnz\" (UID: \"7a12fe6d-d25a-4b08-90da-cc9e510980a7\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-chmnz" Dec 03 00:07:30 crc kubenswrapper[4953]: I1203 00:07:30.073358 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/7a12fe6d-d25a-4b08-90da-cc9e510980a7-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-chmnz\" (UID: \"7a12fe6d-d25a-4b08-90da-cc9e510980a7\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-chmnz" Dec 03 00:07:30 crc kubenswrapper[4953]: I1203 00:07:30.090447 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f94f8c51-c665-4ca3-bf17-7a04ea5c7ade\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72c5cd7dc1c3a8adfedab6ba43a59d80b3efe65f2bb56be95e76cc86a409dcf3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2810ffc1b141fa7da811406053db9d72c708edf5f2bb1dd75e4d52ab35fd24c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c67b4277b20a9f1287f50ef82d0785c9d02543a8e7ed26e0413eae332b866db7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://258d78e1ad9d3dcda6976909ae3fe71aecead5b995fc66b3fd6e2722d80b5ab8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:06:51Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:30Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:30 crc kubenswrapper[4953]: I1203 00:07:30.110049 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09657a9f448a4c92a759c73a0a892076d7eb46393e6fd565e5f73f34c71aac23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5af5eac4a58870f63497f99b503cfa5fb922278095e0332ec50376925ecf2109\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:30Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:30 crc kubenswrapper[4953]: I1203 00:07:30.132209 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-wj6rz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"96aa67a0-15dd-420d-a1d5-ed239d7df1fb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01815cfd3f58c3f3f5f2a894c1790e0b69c5060c48dc0b1deff050b6ad7af916\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kf62m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b12bfb5bdf9fd4f2c50a335fc82df6ee0440eaf8741897da03aa2dae368cc6b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b12bfb5bdf9fd4f2c50a335fc82df6ee0440eaf8741897da03aa2dae368cc6b4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:07:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kf62m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83aaa5fa28110bc0ef05aff75cd1442bc9bbdd33b822ae2ef22d4e4ab83238bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://83aaa5fa28110bc0ef05aff75cd1442bc9bbdd33b822ae2ef22d4e4ab83238bc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:07:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:07:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kf62m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ab6213a33a19fedebce6d74a0ebbb3b3a2de5687504e21cc24093cc42ade635\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ab6213a33a19fedebce6d74a0ebbb3b3a2de5687504e21cc24093cc42ade635\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:07:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:07:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kf62m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://de480488843afa6b8dede1e1b42b5181dca1993e2f39dc1b461a18a57e3668e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://de480488843afa6b8dede1e1b42b5181dca1993e2f39dc1b461a18a57e3668e1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:07:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kf62m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://97be4ba692bc61078d4b0ba39fb0c12772548c6c9251ff3356ea702f151223cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://97be4ba692bc61078d4b0ba39fb0c12772548c6c9251ff3356ea702f151223cd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:07:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kf62m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1a1f009fcbd8af9217db1dc97f662719f8c1db232df0e2804e44c881abb1df2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b1a1f009fcbd8af9217db1dc97f662719f8c1db232df0e2804e44c881abb1df2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:07:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:07:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kf62m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:07:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-wj6rz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:30Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:30 crc kubenswrapper[4953]: I1203 00:07:30.147494 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-kl5s9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d85e5c97-16ad-4bd7-9e73-b7f12678be97\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c21df91a9e4e3165686882b2eb559498224063dc633c3a63f958598108eca29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j24lb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:07:16Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-kl5s9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:30Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:30 crc kubenswrapper[4953]: I1203 00:07:30.151789 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:30 crc kubenswrapper[4953]: I1203 00:07:30.151858 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:30 crc kubenswrapper[4953]: I1203 00:07:30.151915 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:30 crc kubenswrapper[4953]: I1203 00:07:30.151946 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:30 crc kubenswrapper[4953]: I1203 00:07:30.151969 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:30Z","lastTransitionTime":"2025-12-03T00:07:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:30 crc kubenswrapper[4953]: I1203 00:07:30.163830 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5cmlc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7e11a1d3-ecee-4bf0-87bf-085fab97b73d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d42390b9c53efc63add8ddcebf03c976f5e5448eaa2c34e076b2dc90dbac60fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6x2sc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:07:18Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5cmlc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:30Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:30 crc kubenswrapper[4953]: I1203 00:07:30.174376 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/7a12fe6d-d25a-4b08-90da-cc9e510980a7-env-overrides\") pod \"ovnkube-control-plane-749d76644c-chmnz\" (UID: \"7a12fe6d-d25a-4b08-90da-cc9e510980a7\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-chmnz" Dec 03 00:07:30 crc kubenswrapper[4953]: I1203 00:07:30.174514 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/7a12fe6d-d25a-4b08-90da-cc9e510980a7-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-chmnz\" (UID: \"7a12fe6d-d25a-4b08-90da-cc9e510980a7\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-chmnz" Dec 03 00:07:30 crc kubenswrapper[4953]: I1203 00:07:30.174577 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c982g\" (UniqueName: \"kubernetes.io/projected/7a12fe6d-d25a-4b08-90da-cc9e510980a7-kube-api-access-c982g\") pod \"ovnkube-control-plane-749d76644c-chmnz\" (UID: \"7a12fe6d-d25a-4b08-90da-cc9e510980a7\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-chmnz" Dec 03 00:07:30 crc kubenswrapper[4953]: I1203 00:07:30.174654 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/7a12fe6d-d25a-4b08-90da-cc9e510980a7-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-chmnz\" (UID: \"7a12fe6d-d25a-4b08-90da-cc9e510980a7\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-chmnz" Dec 03 00:07:30 crc kubenswrapper[4953]: I1203 00:07:30.175749 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/7a12fe6d-d25a-4b08-90da-cc9e510980a7-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-chmnz\" (UID: \"7a12fe6d-d25a-4b08-90da-cc9e510980a7\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-chmnz" Dec 03 00:07:30 crc kubenswrapper[4953]: I1203 00:07:30.176434 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/7a12fe6d-d25a-4b08-90da-cc9e510980a7-env-overrides\") pod \"ovnkube-control-plane-749d76644c-chmnz\" (UID: \"7a12fe6d-d25a-4b08-90da-cc9e510980a7\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-chmnz" Dec 03 00:07:30 crc kubenswrapper[4953]: I1203 00:07:30.180781 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-chmnz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a12fe6d-d25a-4b08-90da-cc9e510980a7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:29Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:29Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c982g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c982g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:07:29Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-chmnz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:30Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:30 crc kubenswrapper[4953]: I1203 00:07:30.182285 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/7a12fe6d-d25a-4b08-90da-cc9e510980a7-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-chmnz\" (UID: \"7a12fe6d-d25a-4b08-90da-cc9e510980a7\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-chmnz" Dec 03 00:07:30 crc kubenswrapper[4953]: I1203 00:07:30.202387 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c982g\" (UniqueName: \"kubernetes.io/projected/7a12fe6d-d25a-4b08-90da-cc9e510980a7-kube-api-access-c982g\") pod \"ovnkube-control-plane-749d76644c-chmnz\" (UID: \"7a12fe6d-d25a-4b08-90da-cc9e510980a7\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-chmnz" Dec 03 00:07:30 crc kubenswrapper[4953]: I1203 00:07:30.206646 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cf821840-5353-4620-8f73-99b56b371ab0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cae5e2994fcc24dc9b2ea73e79572615b3311edc42642b7f278fb84123989b23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://29e4b9d1b70a93fb133e0759ad148ae133766a66efeae3ff06a21dfae0e805ff\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dec43536772aa7c3f5a662fa7f3bd73b1cae2791ad454f85792287828d0ffeed\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://034e0bac6892d361700b8f33d6427b8a55e2bd3f89b14ea77a374c6be94d25e2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://761973393a93c59c23e69f41b8ac0d16abc35abd61ecbc6d51a42d94cd664233\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T00:07:09Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 00:07:03.713374 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 00:07:03.715052 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1528563048/tls.crt::/tmp/serving-cert-1528563048/tls.key\\\\\\\"\\\\nI1203 00:07:09.300052 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 00:07:09.304192 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 00:07:09.304255 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 00:07:09.304309 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 00:07:09.304323 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 00:07:09.313563 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 00:07:09.313597 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 00:07:09.313603 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 00:07:09.313609 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 00:07:09.313613 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 00:07:09.313618 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 00:07:09.313622 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 00:07:09.313709 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 00:07:09.327347 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T00:06:53Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c143d4232fb8bf7ef5999a664e431a00ab0db030f4d31a28d0bb21e00e38581\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:53Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72af6c27296eae2e275d455795c68f0d272869d63642b8353b17d466d0c901cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://72af6c27296eae2e275d455795c68f0d272869d63642b8353b17d466d0c901cf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:06:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:06:51Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:30Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:30 crc kubenswrapper[4953]: I1203 00:07:30.225065 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:10Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:30Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:30 crc kubenswrapper[4953]: I1203 00:07:30.245909 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce8b414e1fc22eb30decb6a6411b672e7f443055cb89ebb029fbf6ffc1d5dca7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:30Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:30 crc kubenswrapper[4953]: I1203 00:07:30.256066 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:30 crc kubenswrapper[4953]: I1203 00:07:30.256153 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:30 crc kubenswrapper[4953]: I1203 00:07:30.256180 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:30 crc kubenswrapper[4953]: I1203 00:07:30.256215 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:30 crc kubenswrapper[4953]: I1203 00:07:30.256235 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:30Z","lastTransitionTime":"2025-12-03T00:07:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:30 crc kubenswrapper[4953]: I1203 00:07:30.262338 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-6dfbn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53ffb1e9-9a2e-4869-80d7-776f4aa7ef8b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ab4205c0d38ad6a0c32077eadb1fd5fd3a279d9c8b8a2ffbe693ec2c4018370c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ct26c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7df84405eda520ad5a22d35750edf07ef3c3977a756ad5739c54f91cba9da7b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ct26c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:07:16Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-6dfbn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:30Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:30 crc kubenswrapper[4953]: I1203 00:07:30.293418 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-txffx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8816ebee-c735-4b8d-b1ec-65349707b347\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:16Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:16Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4b85224419a4075acb340e4b11095f2adb5c316e285afa4e19295c2181c0d24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0f6f645a58993f052a6f365bb808f62d99f7fea74886c5cc022b2d86112321a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a61b951a9cbab2c48c0f2df8351f97d7ff71bf3674f641fe5eac30155232b224\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd5f428de229749485267e81d34e31b708728b52474032e88460e54747f21e11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://00a2ea4a9bb870fce30e7fa845f787f184e7a19e42c95511655ea8c5f84a26d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc0a5d4cc9d614d8492626a1b114bdc52f510eb5965d842c02c083f08d4849a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://efbfc74a8598c47a658c06b597fa36d2a5ac8d02265d2864fd2832dd11ecb2cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aecc0ae6dd34893c9023da3b94ca01ae96ac1e7cd7f5fd18748cce6c273a7f9e\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T00:07:26Z\\\",\\\"message\\\":\\\"ller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1203 00:07:25.992760 6257 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1203 00:07:25.992789 6257 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1203 00:07:25.992837 6257 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1203 00:07:25.992889 6257 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1203 00:07:25.992927 6257 handler.go:208] Removed *v1.Node event handler 7\\\\nI1203 00:07:25.992967 6257 handler.go:208] Removed *v1.Node event handler 2\\\\nI1203 00:07:25.992993 6257 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1203 00:07:25.992947 6257 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1203 00:07:25.993079 6257 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1203 00:07:25.992969 6257 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1203 00:07:25.993139 6257 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1203 00:07:25.993168 6257 factory.go:656] Stopping watch factory\\\\nI1203 00:07:25.993190 6257 ovnkube.go:599] Stopped ovnkube\\\\nI1203 00:07:25.993215 6257 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1203 00:07:25.993246 6257 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T00:07:23Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7deadc8e135ff1bc7c5a2f093e489bb5b4b6f166bc874c6acc2a841c5397f11e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af4c7764fca87144d01f3738ea010308744ccf50c54b7c471ccc213112de00d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af4c7764fca87144d01f3738ea010308744ccf50c54b7c471ccc213112de00d5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:07:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:07:16Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-txffx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:30Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:30 crc kubenswrapper[4953]: I1203 00:07:30.312437 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-chmnz" Dec 03 00:07:30 crc kubenswrapper[4953]: W1203 00:07:30.333469 4953 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7a12fe6d_d25a_4b08_90da_cc9e510980a7.slice/crio-53842eec8a860157c58f1d570c0a95a0d31bee2cccdf4a273fa434111931b6fd WatchSource:0}: Error finding container 53842eec8a860157c58f1d570c0a95a0d31bee2cccdf4a273fa434111931b6fd: Status 404 returned error can't find the container with id 53842eec8a860157c58f1d570c0a95a0d31bee2cccdf4a273fa434111931b6fd Dec 03 00:07:30 crc kubenswrapper[4953]: I1203 00:07:30.347413 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 00:07:30 crc kubenswrapper[4953]: I1203 00:07:30.347464 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 00:07:30 crc kubenswrapper[4953]: I1203 00:07:30.347422 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 00:07:30 crc kubenswrapper[4953]: E1203 00:07:30.347613 4953 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 00:07:30 crc kubenswrapper[4953]: E1203 00:07:30.347727 4953 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 00:07:30 crc kubenswrapper[4953]: E1203 00:07:30.347870 4953 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 00:07:30 crc kubenswrapper[4953]: I1203 00:07:30.360477 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:30 crc kubenswrapper[4953]: I1203 00:07:30.360522 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:30 crc kubenswrapper[4953]: I1203 00:07:30.360542 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:30 crc kubenswrapper[4953]: I1203 00:07:30.360565 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:30 crc kubenswrapper[4953]: I1203 00:07:30.360582 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:30Z","lastTransitionTime":"2025-12-03T00:07:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:30 crc kubenswrapper[4953]: I1203 00:07:30.463487 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:30 crc kubenswrapper[4953]: I1203 00:07:30.463525 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:30 crc kubenswrapper[4953]: I1203 00:07:30.463542 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:30 crc kubenswrapper[4953]: I1203 00:07:30.463562 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:30 crc kubenswrapper[4953]: I1203 00:07:30.463577 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:30Z","lastTransitionTime":"2025-12-03T00:07:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:30 crc kubenswrapper[4953]: I1203 00:07:30.566839 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:30 crc kubenswrapper[4953]: I1203 00:07:30.566904 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:30 crc kubenswrapper[4953]: I1203 00:07:30.566920 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:30 crc kubenswrapper[4953]: I1203 00:07:30.566945 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:30 crc kubenswrapper[4953]: I1203 00:07:30.566962 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:30Z","lastTransitionTime":"2025-12-03T00:07:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:30 crc kubenswrapper[4953]: I1203 00:07:30.666682 4953 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-txffx_8816ebee-c735-4b8d-b1ec-65349707b347/ovnkube-controller/1.log" Dec 03 00:07:30 crc kubenswrapper[4953]: I1203 00:07:30.667545 4953 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-txffx_8816ebee-c735-4b8d-b1ec-65349707b347/ovnkube-controller/0.log" Dec 03 00:07:30 crc kubenswrapper[4953]: I1203 00:07:30.669808 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:30 crc kubenswrapper[4953]: I1203 00:07:30.669849 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:30 crc kubenswrapper[4953]: I1203 00:07:30.669865 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:30 crc kubenswrapper[4953]: I1203 00:07:30.669888 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:30 crc kubenswrapper[4953]: I1203 00:07:30.669908 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:30Z","lastTransitionTime":"2025-12-03T00:07:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:30 crc kubenswrapper[4953]: I1203 00:07:30.671780 4953 generic.go:334] "Generic (PLEG): container finished" podID="8816ebee-c735-4b8d-b1ec-65349707b347" containerID="efbfc74a8598c47a658c06b597fa36d2a5ac8d02265d2864fd2832dd11ecb2cc" exitCode=1 Dec 03 00:07:30 crc kubenswrapper[4953]: I1203 00:07:30.671868 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-txffx" event={"ID":"8816ebee-c735-4b8d-b1ec-65349707b347","Type":"ContainerDied","Data":"efbfc74a8598c47a658c06b597fa36d2a5ac8d02265d2864fd2832dd11ecb2cc"} Dec 03 00:07:30 crc kubenswrapper[4953]: I1203 00:07:30.672010 4953 scope.go:117] "RemoveContainer" containerID="aecc0ae6dd34893c9023da3b94ca01ae96ac1e7cd7f5fd18748cce6c273a7f9e" Dec 03 00:07:30 crc kubenswrapper[4953]: I1203 00:07:30.672898 4953 scope.go:117] "RemoveContainer" containerID="efbfc74a8598c47a658c06b597fa36d2a5ac8d02265d2864fd2832dd11ecb2cc" Dec 03 00:07:30 crc kubenswrapper[4953]: E1203 00:07:30.673166 4953 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-txffx_openshift-ovn-kubernetes(8816ebee-c735-4b8d-b1ec-65349707b347)\"" pod="openshift-ovn-kubernetes/ovnkube-node-txffx" podUID="8816ebee-c735-4b8d-b1ec-65349707b347" Dec 03 00:07:30 crc kubenswrapper[4953]: I1203 00:07:30.676322 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-chmnz" event={"ID":"7a12fe6d-d25a-4b08-90da-cc9e510980a7","Type":"ContainerStarted","Data":"711ab839cb9c0a90554dadbd45d810fc78aac09739340482fef8de8743b77283"} Dec 03 00:07:30 crc kubenswrapper[4953]: I1203 00:07:30.676365 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-chmnz" event={"ID":"7a12fe6d-d25a-4b08-90da-cc9e510980a7","Type":"ContainerStarted","Data":"53842eec8a860157c58f1d570c0a95a0d31bee2cccdf4a273fa434111931b6fd"} Dec 03 00:07:30 crc kubenswrapper[4953]: I1203 00:07:30.695114 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-wj6rz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"96aa67a0-15dd-420d-a1d5-ed239d7df1fb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01815cfd3f58c3f3f5f2a894c1790e0b69c5060c48dc0b1deff050b6ad7af916\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kf62m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b12bfb5bdf9fd4f2c50a335fc82df6ee0440eaf8741897da03aa2dae368cc6b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b12bfb5bdf9fd4f2c50a335fc82df6ee0440eaf8741897da03aa2dae368cc6b4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:07:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kf62m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83aaa5fa28110bc0ef05aff75cd1442bc9bbdd33b822ae2ef22d4e4ab83238bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://83aaa5fa28110bc0ef05aff75cd1442bc9bbdd33b822ae2ef22d4e4ab83238bc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:07:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:07:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kf62m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ab6213a33a19fedebce6d74a0ebbb3b3a2de5687504e21cc24093cc42ade635\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ab6213a33a19fedebce6d74a0ebbb3b3a2de5687504e21cc24093cc42ade635\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:07:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:07:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kf62m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://de480488843afa6b8dede1e1b42b5181dca1993e2f39dc1b461a18a57e3668e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://de480488843afa6b8dede1e1b42b5181dca1993e2f39dc1b461a18a57e3668e1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:07:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kf62m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://97be4ba692bc61078d4b0ba39fb0c12772548c6c9251ff3356ea702f151223cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://97be4ba692bc61078d4b0ba39fb0c12772548c6c9251ff3356ea702f151223cd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:07:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kf62m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1a1f009fcbd8af9217db1dc97f662719f8c1db232df0e2804e44c881abb1df2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b1a1f009fcbd8af9217db1dc97f662719f8c1db232df0e2804e44c881abb1df2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:07:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:07:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kf62m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:07:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-wj6rz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:30Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:30 crc kubenswrapper[4953]: I1203 00:07:30.709614 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-kl5s9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d85e5c97-16ad-4bd7-9e73-b7f12678be97\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c21df91a9e4e3165686882b2eb559498224063dc633c3a63f958598108eca29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j24lb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:07:16Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-kl5s9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:30Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:30 crc kubenswrapper[4953]: I1203 00:07:30.724413 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f94f8c51-c665-4ca3-bf17-7a04ea5c7ade\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72c5cd7dc1c3a8adfedab6ba43a59d80b3efe65f2bb56be95e76cc86a409dcf3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2810ffc1b141fa7da811406053db9d72c708edf5f2bb1dd75e4d52ab35fd24c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c67b4277b20a9f1287f50ef82d0785c9d02543a8e7ed26e0413eae332b866db7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://258d78e1ad9d3dcda6976909ae3fe71aecead5b995fc66b3fd6e2722d80b5ab8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:06:51Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:30Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:30 crc kubenswrapper[4953]: I1203 00:07:30.738262 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09657a9f448a4c92a759c73a0a892076d7eb46393e6fd565e5f73f34c71aac23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5af5eac4a58870f63497f99b503cfa5fb922278095e0332ec50376925ecf2109\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:30Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:30 crc kubenswrapper[4953]: I1203 00:07:30.751145 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-6dfbn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53ffb1e9-9a2e-4869-80d7-776f4aa7ef8b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ab4205c0d38ad6a0c32077eadb1fd5fd3a279d9c8b8a2ffbe693ec2c4018370c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ct26c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7df84405eda520ad5a22d35750edf07ef3c3977a756ad5739c54f91cba9da7b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ct26c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:07:16Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-6dfbn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:30Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:30 crc kubenswrapper[4953]: I1203 00:07:30.772215 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:30 crc kubenswrapper[4953]: I1203 00:07:30.772269 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:30 crc kubenswrapper[4953]: I1203 00:07:30.772287 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:30 crc kubenswrapper[4953]: I1203 00:07:30.772306 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:30 crc kubenswrapper[4953]: I1203 00:07:30.772321 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:30Z","lastTransitionTime":"2025-12-03T00:07:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:30 crc kubenswrapper[4953]: I1203 00:07:30.777729 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-txffx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8816ebee-c735-4b8d-b1ec-65349707b347\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:16Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:16Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4b85224419a4075acb340e4b11095f2adb5c316e285afa4e19295c2181c0d24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0f6f645a58993f052a6f365bb808f62d99f7fea74886c5cc022b2d86112321a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a61b951a9cbab2c48c0f2df8351f97d7ff71bf3674f641fe5eac30155232b224\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd5f428de229749485267e81d34e31b708728b52474032e88460e54747f21e11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://00a2ea4a9bb870fce30e7fa845f787f184e7a19e42c95511655ea8c5f84a26d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc0a5d4cc9d614d8492626a1b114bdc52f510eb5965d842c02c083f08d4849a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://efbfc74a8598c47a658c06b597fa36d2a5ac8d02265d2864fd2832dd11ecb2cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aecc0ae6dd34893c9023da3b94ca01ae96ac1e7cd7f5fd18748cce6c273a7f9e\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T00:07:26Z\\\",\\\"message\\\":\\\"ller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1203 00:07:25.992760 6257 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1203 00:07:25.992789 6257 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1203 00:07:25.992837 6257 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1203 00:07:25.992889 6257 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1203 00:07:25.992927 6257 handler.go:208] Removed *v1.Node event handler 7\\\\nI1203 00:07:25.992967 6257 handler.go:208] Removed *v1.Node event handler 2\\\\nI1203 00:07:25.992993 6257 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1203 00:07:25.992947 6257 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1203 00:07:25.993079 6257 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1203 00:07:25.992969 6257 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1203 00:07:25.993139 6257 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1203 00:07:25.993168 6257 factory.go:656] Stopping watch factory\\\\nI1203 00:07:25.993190 6257 ovnkube.go:599] Stopped ovnkube\\\\nI1203 00:07:25.993215 6257 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1203 00:07:25.993246 6257 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T00:07:23Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://efbfc74a8598c47a658c06b597fa36d2a5ac8d02265d2864fd2832dd11ecb2cc\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T00:07:29Z\\\",\\\"message\\\":\\\" 6400 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1203 00:07:29.614806 6400 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1203 00:07:29.614821 6400 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1203 00:07:29.614822 6400 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 00:07:29.614854 6400 factory.go:656] Stopping watch factory\\\\nI1203 00:07:29.614869 6400 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1203 00:07:29.614878 6400 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1203 00:07:29.614875 6400 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1203 00:07:29.614894 6400 handler.go:208] Removed *v1.Node event handler 7\\\\nI1203 00:07:29.614906 6400 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1203 00:07:29.614915 6400 reflector.go:311] Stopping reflector *v1.NetworkAttachmentDefinition (0s) from github.com/k8snetworkplumbingwg/network-attachment-definition-client/pkg/client/informers/externalversions/factory.go:117\\\\nI1203 00:07:29.615144 6400 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 00:07:29.615318 6400 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T00:07:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7deadc8e135ff1bc7c5a2f093e489bb5b4b6f166bc874c6acc2a841c5397f11e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af4c7764fca87144d01f3738ea010308744ccf50c54b7c471ccc213112de00d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af4c7764fca87144d01f3738ea010308744ccf50c54b7c471ccc213112de00d5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:07:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:07:16Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-txffx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:30Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:30 crc kubenswrapper[4953]: I1203 00:07:30.791503 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5cmlc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7e11a1d3-ecee-4bf0-87bf-085fab97b73d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d42390b9c53efc63add8ddcebf03c976f5e5448eaa2c34e076b2dc90dbac60fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6x2sc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:07:18Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5cmlc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:30Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:30 crc kubenswrapper[4953]: I1203 00:07:30.804365 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-chmnz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a12fe6d-d25a-4b08-90da-cc9e510980a7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:29Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:29Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c982g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c982g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:07:29Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-chmnz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:30Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:30 crc kubenswrapper[4953]: I1203 00:07:30.819848 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cf821840-5353-4620-8f73-99b56b371ab0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cae5e2994fcc24dc9b2ea73e79572615b3311edc42642b7f278fb84123989b23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://29e4b9d1b70a93fb133e0759ad148ae133766a66efeae3ff06a21dfae0e805ff\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dec43536772aa7c3f5a662fa7f3bd73b1cae2791ad454f85792287828d0ffeed\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://034e0bac6892d361700b8f33d6427b8a55e2bd3f89b14ea77a374c6be94d25e2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://761973393a93c59c23e69f41b8ac0d16abc35abd61ecbc6d51a42d94cd664233\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T00:07:09Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 00:07:03.713374 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 00:07:03.715052 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1528563048/tls.crt::/tmp/serving-cert-1528563048/tls.key\\\\\\\"\\\\nI1203 00:07:09.300052 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 00:07:09.304192 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 00:07:09.304255 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 00:07:09.304309 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 00:07:09.304323 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 00:07:09.313563 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 00:07:09.313597 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 00:07:09.313603 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 00:07:09.313609 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 00:07:09.313613 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 00:07:09.313618 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 00:07:09.313622 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 00:07:09.313709 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 00:07:09.327347 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T00:06:53Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c143d4232fb8bf7ef5999a664e431a00ab0db030f4d31a28d0bb21e00e38581\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:53Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72af6c27296eae2e275d455795c68f0d272869d63642b8353b17d466d0c901cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://72af6c27296eae2e275d455795c68f0d272869d63642b8353b17d466d0c901cf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:06:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:06:51Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:30Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:30 crc kubenswrapper[4953]: I1203 00:07:30.837362 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:10Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:30Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:30 crc kubenswrapper[4953]: I1203 00:07:30.854648 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce8b414e1fc22eb30decb6a6411b672e7f443055cb89ebb029fbf6ffc1d5dca7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:30Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:30 crc kubenswrapper[4953]: I1203 00:07:30.868266 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2ed9c7fa96c8951fbc1728fe7a91fefdd999077285621df9c1abedf9ae0c449\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:30Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:30 crc kubenswrapper[4953]: I1203 00:07:30.874484 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:30 crc kubenswrapper[4953]: I1203 00:07:30.874736 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:30 crc kubenswrapper[4953]: I1203 00:07:30.874838 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:30 crc kubenswrapper[4953]: I1203 00:07:30.874928 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:30 crc kubenswrapper[4953]: I1203 00:07:30.875060 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:30Z","lastTransitionTime":"2025-12-03T00:07:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:30 crc kubenswrapper[4953]: I1203 00:07:30.882084 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:30Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:30 crc kubenswrapper[4953]: I1203 00:07:30.893663 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:10Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:30Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:30 crc kubenswrapper[4953]: I1203 00:07:30.912152 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-675qk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"43a4bad8-71f4-441a-8302-aeb526a986af\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5b5794fef6477aab5deed8f642e8cc50b37eb1fd0cc6f19c003f34ad31c1e3c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gdrw4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:07:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-675qk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:30Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:30 crc kubenswrapper[4953]: I1203 00:07:30.977499 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:30 crc kubenswrapper[4953]: I1203 00:07:30.978104 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:30 crc kubenswrapper[4953]: I1203 00:07:30.978135 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:30 crc kubenswrapper[4953]: I1203 00:07:30.978156 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:30 crc kubenswrapper[4953]: I1203 00:07:30.978170 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:30Z","lastTransitionTime":"2025-12-03T00:07:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:31 crc kubenswrapper[4953]: I1203 00:07:31.081620 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:31 crc kubenswrapper[4953]: I1203 00:07:31.081669 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:31 crc kubenswrapper[4953]: I1203 00:07:31.081678 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:31 crc kubenswrapper[4953]: I1203 00:07:31.081694 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:31 crc kubenswrapper[4953]: I1203 00:07:31.081708 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:31Z","lastTransitionTime":"2025-12-03T00:07:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:31 crc kubenswrapper[4953]: I1203 00:07:31.120856 4953 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/network-metrics-daemon-xshk4"] Dec 03 00:07:31 crc kubenswrapper[4953]: I1203 00:07:31.121598 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xshk4" Dec 03 00:07:31 crc kubenswrapper[4953]: E1203 00:07:31.121699 4953 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xshk4" podUID="bd03a758-0ca9-47d8-86f7-18b2d610df43" Dec 03 00:07:31 crc kubenswrapper[4953]: I1203 00:07:31.141293 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:31Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:31 crc kubenswrapper[4953]: I1203 00:07:31.157953 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2ed9c7fa96c8951fbc1728fe7a91fefdd999077285621df9c1abedf9ae0c449\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:31Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:31 crc kubenswrapper[4953]: I1203 00:07:31.178768 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:10Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:31Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:31 crc kubenswrapper[4953]: I1203 00:07:31.185699 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-47gfh\" (UniqueName: \"kubernetes.io/projected/bd03a758-0ca9-47d8-86f7-18b2d610df43-kube-api-access-47gfh\") pod \"network-metrics-daemon-xshk4\" (UID: \"bd03a758-0ca9-47d8-86f7-18b2d610df43\") " pod="openshift-multus/network-metrics-daemon-xshk4" Dec 03 00:07:31 crc kubenswrapper[4953]: I1203 00:07:31.185921 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/bd03a758-0ca9-47d8-86f7-18b2d610df43-metrics-certs\") pod \"network-metrics-daemon-xshk4\" (UID: \"bd03a758-0ca9-47d8-86f7-18b2d610df43\") " pod="openshift-multus/network-metrics-daemon-xshk4" Dec 03 00:07:31 crc kubenswrapper[4953]: I1203 00:07:31.187018 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:31 crc kubenswrapper[4953]: I1203 00:07:31.187448 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:31 crc kubenswrapper[4953]: I1203 00:07:31.187468 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:31 crc kubenswrapper[4953]: I1203 00:07:31.187491 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:31 crc kubenswrapper[4953]: I1203 00:07:31.187508 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:31Z","lastTransitionTime":"2025-12-03T00:07:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:31 crc kubenswrapper[4953]: I1203 00:07:31.199404 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-675qk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"43a4bad8-71f4-441a-8302-aeb526a986af\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5b5794fef6477aab5deed8f642e8cc50b37eb1fd0cc6f19c003f34ad31c1e3c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gdrw4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:07:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-675qk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:31Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:31 crc kubenswrapper[4953]: I1203 00:07:31.217284 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09657a9f448a4c92a759c73a0a892076d7eb46393e6fd565e5f73f34c71aac23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5af5eac4a58870f63497f99b503cfa5fb922278095e0332ec50376925ecf2109\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:31Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:31 crc kubenswrapper[4953]: I1203 00:07:31.236647 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-wj6rz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"96aa67a0-15dd-420d-a1d5-ed239d7df1fb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01815cfd3f58c3f3f5f2a894c1790e0b69c5060c48dc0b1deff050b6ad7af916\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kf62m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b12bfb5bdf9fd4f2c50a335fc82df6ee0440eaf8741897da03aa2dae368cc6b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b12bfb5bdf9fd4f2c50a335fc82df6ee0440eaf8741897da03aa2dae368cc6b4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:07:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kf62m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83aaa5fa28110bc0ef05aff75cd1442bc9bbdd33b822ae2ef22d4e4ab83238bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://83aaa5fa28110bc0ef05aff75cd1442bc9bbdd33b822ae2ef22d4e4ab83238bc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:07:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:07:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kf62m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ab6213a33a19fedebce6d74a0ebbb3b3a2de5687504e21cc24093cc42ade635\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ab6213a33a19fedebce6d74a0ebbb3b3a2de5687504e21cc24093cc42ade635\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:07:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:07:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kf62m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://de480488843afa6b8dede1e1b42b5181dca1993e2f39dc1b461a18a57e3668e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://de480488843afa6b8dede1e1b42b5181dca1993e2f39dc1b461a18a57e3668e1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:07:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kf62m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://97be4ba692bc61078d4b0ba39fb0c12772548c6c9251ff3356ea702f151223cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://97be4ba692bc61078d4b0ba39fb0c12772548c6c9251ff3356ea702f151223cd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:07:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kf62m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1a1f009fcbd8af9217db1dc97f662719f8c1db232df0e2804e44c881abb1df2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b1a1f009fcbd8af9217db1dc97f662719f8c1db232df0e2804e44c881abb1df2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:07:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:07:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kf62m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:07:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-wj6rz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:31Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:31 crc kubenswrapper[4953]: I1203 00:07:31.255191 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-kl5s9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d85e5c97-16ad-4bd7-9e73-b7f12678be97\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c21df91a9e4e3165686882b2eb559498224063dc633c3a63f958598108eca29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j24lb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:07:16Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-kl5s9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:31Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:31 crc kubenswrapper[4953]: I1203 00:07:31.275968 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f94f8c51-c665-4ca3-bf17-7a04ea5c7ade\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72c5cd7dc1c3a8adfedab6ba43a59d80b3efe65f2bb56be95e76cc86a409dcf3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2810ffc1b141fa7da811406053db9d72c708edf5f2bb1dd75e4d52ab35fd24c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c67b4277b20a9f1287f50ef82d0785c9d02543a8e7ed26e0413eae332b866db7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://258d78e1ad9d3dcda6976909ae3fe71aecead5b995fc66b3fd6e2722d80b5ab8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:06:51Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:31Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:31 crc kubenswrapper[4953]: I1203 00:07:31.287050 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-47gfh\" (UniqueName: \"kubernetes.io/projected/bd03a758-0ca9-47d8-86f7-18b2d610df43-kube-api-access-47gfh\") pod \"network-metrics-daemon-xshk4\" (UID: \"bd03a758-0ca9-47d8-86f7-18b2d610df43\") " pod="openshift-multus/network-metrics-daemon-xshk4" Dec 03 00:07:31 crc kubenswrapper[4953]: I1203 00:07:31.287142 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/bd03a758-0ca9-47d8-86f7-18b2d610df43-metrics-certs\") pod \"network-metrics-daemon-xshk4\" (UID: \"bd03a758-0ca9-47d8-86f7-18b2d610df43\") " pod="openshift-multus/network-metrics-daemon-xshk4" Dec 03 00:07:31 crc kubenswrapper[4953]: E1203 00:07:31.287391 4953 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 03 00:07:31 crc kubenswrapper[4953]: E1203 00:07:31.287501 4953 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/bd03a758-0ca9-47d8-86f7-18b2d610df43-metrics-certs podName:bd03a758-0ca9-47d8-86f7-18b2d610df43 nodeName:}" failed. No retries permitted until 2025-12-03 00:07:31.787469069 +0000 UTC m=+40.771749892 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/bd03a758-0ca9-47d8-86f7-18b2d610df43-metrics-certs") pod "network-metrics-daemon-xshk4" (UID: "bd03a758-0ca9-47d8-86f7-18b2d610df43") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 03 00:07:31 crc kubenswrapper[4953]: I1203 00:07:31.289802 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:31 crc kubenswrapper[4953]: I1203 00:07:31.289904 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:31 crc kubenswrapper[4953]: I1203 00:07:31.289932 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:31 crc kubenswrapper[4953]: I1203 00:07:31.289962 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:31 crc kubenswrapper[4953]: I1203 00:07:31.290016 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:31Z","lastTransitionTime":"2025-12-03T00:07:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:31 crc kubenswrapper[4953]: I1203 00:07:31.294908 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce8b414e1fc22eb30decb6a6411b672e7f443055cb89ebb029fbf6ffc1d5dca7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:31Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:31 crc kubenswrapper[4953]: I1203 00:07:31.309017 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-6dfbn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53ffb1e9-9a2e-4869-80d7-776f4aa7ef8b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ab4205c0d38ad6a0c32077eadb1fd5fd3a279d9c8b8a2ffbe693ec2c4018370c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ct26c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7df84405eda520ad5a22d35750edf07ef3c3977a756ad5739c54f91cba9da7b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ct26c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:07:16Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-6dfbn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:31Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:31 crc kubenswrapper[4953]: I1203 00:07:31.317187 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-47gfh\" (UniqueName: \"kubernetes.io/projected/bd03a758-0ca9-47d8-86f7-18b2d610df43-kube-api-access-47gfh\") pod \"network-metrics-daemon-xshk4\" (UID: \"bd03a758-0ca9-47d8-86f7-18b2d610df43\") " pod="openshift-multus/network-metrics-daemon-xshk4" Dec 03 00:07:31 crc kubenswrapper[4953]: I1203 00:07:31.336670 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-txffx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8816ebee-c735-4b8d-b1ec-65349707b347\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:16Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:16Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4b85224419a4075acb340e4b11095f2adb5c316e285afa4e19295c2181c0d24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0f6f645a58993f052a6f365bb808f62d99f7fea74886c5cc022b2d86112321a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a61b951a9cbab2c48c0f2df8351f97d7ff71bf3674f641fe5eac30155232b224\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd5f428de229749485267e81d34e31b708728b52474032e88460e54747f21e11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://00a2ea4a9bb870fce30e7fa845f787f184e7a19e42c95511655ea8c5f84a26d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc0a5d4cc9d614d8492626a1b114bdc52f510eb5965d842c02c083f08d4849a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://efbfc74a8598c47a658c06b597fa36d2a5ac8d02265d2864fd2832dd11ecb2cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aecc0ae6dd34893c9023da3b94ca01ae96ac1e7cd7f5fd18748cce6c273a7f9e\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T00:07:26Z\\\",\\\"message\\\":\\\"ller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1203 00:07:25.992760 6257 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1203 00:07:25.992789 6257 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1203 00:07:25.992837 6257 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1203 00:07:25.992889 6257 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1203 00:07:25.992927 6257 handler.go:208] Removed *v1.Node event handler 7\\\\nI1203 00:07:25.992967 6257 handler.go:208] Removed *v1.Node event handler 2\\\\nI1203 00:07:25.992993 6257 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1203 00:07:25.992947 6257 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1203 00:07:25.993079 6257 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1203 00:07:25.992969 6257 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1203 00:07:25.993139 6257 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1203 00:07:25.993168 6257 factory.go:656] Stopping watch factory\\\\nI1203 00:07:25.993190 6257 ovnkube.go:599] Stopped ovnkube\\\\nI1203 00:07:25.993215 6257 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1203 00:07:25.993246 6257 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T00:07:23Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://efbfc74a8598c47a658c06b597fa36d2a5ac8d02265d2864fd2832dd11ecb2cc\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T00:07:29Z\\\",\\\"message\\\":\\\" 6400 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1203 00:07:29.614806 6400 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1203 00:07:29.614821 6400 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1203 00:07:29.614822 6400 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 00:07:29.614854 6400 factory.go:656] Stopping watch factory\\\\nI1203 00:07:29.614869 6400 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1203 00:07:29.614878 6400 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1203 00:07:29.614875 6400 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1203 00:07:29.614894 6400 handler.go:208] Removed *v1.Node event handler 7\\\\nI1203 00:07:29.614906 6400 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1203 00:07:29.614915 6400 reflector.go:311] Stopping reflector *v1.NetworkAttachmentDefinition (0s) from github.com/k8snetworkplumbingwg/network-attachment-definition-client/pkg/client/informers/externalversions/factory.go:117\\\\nI1203 00:07:29.615144 6400 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 00:07:29.615318 6400 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T00:07:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7deadc8e135ff1bc7c5a2f093e489bb5b4b6f166bc874c6acc2a841c5397f11e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af4c7764fca87144d01f3738ea010308744ccf50c54b7c471ccc213112de00d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af4c7764fca87144d01f3738ea010308744ccf50c54b7c471ccc213112de00d5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:07:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:07:16Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-txffx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:31Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:31 crc kubenswrapper[4953]: I1203 00:07:31.347744 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5cmlc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7e11a1d3-ecee-4bf0-87bf-085fab97b73d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d42390b9c53efc63add8ddcebf03c976f5e5448eaa2c34e076b2dc90dbac60fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6x2sc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:07:18Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5cmlc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:31Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:31 crc kubenswrapper[4953]: I1203 00:07:31.362806 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-chmnz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a12fe6d-d25a-4b08-90da-cc9e510980a7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:29Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:29Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c982g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c982g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:07:29Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-chmnz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:31Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:31 crc kubenswrapper[4953]: I1203 00:07:31.375849 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-xshk4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bd03a758-0ca9-47d8-86f7-18b2d610df43\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47gfh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47gfh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:07:31Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-xshk4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:31Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:31 crc kubenswrapper[4953]: I1203 00:07:31.389725 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cf821840-5353-4620-8f73-99b56b371ab0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cae5e2994fcc24dc9b2ea73e79572615b3311edc42642b7f278fb84123989b23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://29e4b9d1b70a93fb133e0759ad148ae133766a66efeae3ff06a21dfae0e805ff\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dec43536772aa7c3f5a662fa7f3bd73b1cae2791ad454f85792287828d0ffeed\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://034e0bac6892d361700b8f33d6427b8a55e2bd3f89b14ea77a374c6be94d25e2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://761973393a93c59c23e69f41b8ac0d16abc35abd61ecbc6d51a42d94cd664233\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T00:07:09Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 00:07:03.713374 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 00:07:03.715052 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1528563048/tls.crt::/tmp/serving-cert-1528563048/tls.key\\\\\\\"\\\\nI1203 00:07:09.300052 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 00:07:09.304192 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 00:07:09.304255 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 00:07:09.304309 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 00:07:09.304323 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 00:07:09.313563 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 00:07:09.313597 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 00:07:09.313603 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 00:07:09.313609 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 00:07:09.313613 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 00:07:09.313618 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 00:07:09.313622 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 00:07:09.313709 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 00:07:09.327347 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T00:06:53Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c143d4232fb8bf7ef5999a664e431a00ab0db030f4d31a28d0bb21e00e38581\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:53Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72af6c27296eae2e275d455795c68f0d272869d63642b8353b17d466d0c901cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://72af6c27296eae2e275d455795c68f0d272869d63642b8353b17d466d0c901cf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:06:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:06:51Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:31Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:31 crc kubenswrapper[4953]: I1203 00:07:31.392818 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:31 crc kubenswrapper[4953]: I1203 00:07:31.392874 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:31 crc kubenswrapper[4953]: I1203 00:07:31.392895 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:31 crc kubenswrapper[4953]: I1203 00:07:31.392918 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:31 crc kubenswrapper[4953]: I1203 00:07:31.392937 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:31Z","lastTransitionTime":"2025-12-03T00:07:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:31 crc kubenswrapper[4953]: I1203 00:07:31.407871 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:10Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:31Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:31 crc kubenswrapper[4953]: I1203 00:07:31.424523 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-kl5s9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d85e5c97-16ad-4bd7-9e73-b7f12678be97\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c21df91a9e4e3165686882b2eb559498224063dc633c3a63f958598108eca29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j24lb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:07:16Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-kl5s9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:31Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:31 crc kubenswrapper[4953]: I1203 00:07:31.448517 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f94f8c51-c665-4ca3-bf17-7a04ea5c7ade\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72c5cd7dc1c3a8adfedab6ba43a59d80b3efe65f2bb56be95e76cc86a409dcf3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2810ffc1b141fa7da811406053db9d72c708edf5f2bb1dd75e4d52ab35fd24c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c67b4277b20a9f1287f50ef82d0785c9d02543a8e7ed26e0413eae332b866db7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://258d78e1ad9d3dcda6976909ae3fe71aecead5b995fc66b3fd6e2722d80b5ab8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:06:51Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:31Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:31 crc kubenswrapper[4953]: I1203 00:07:31.475625 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09657a9f448a4c92a759c73a0a892076d7eb46393e6fd565e5f73f34c71aac23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5af5eac4a58870f63497f99b503cfa5fb922278095e0332ec50376925ecf2109\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:31Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:31 crc kubenswrapper[4953]: I1203 00:07:31.494673 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:31 crc kubenswrapper[4953]: I1203 00:07:31.494727 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:31 crc kubenswrapper[4953]: I1203 00:07:31.494745 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:31 crc kubenswrapper[4953]: I1203 00:07:31.494765 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:31 crc kubenswrapper[4953]: I1203 00:07:31.494779 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:31Z","lastTransitionTime":"2025-12-03T00:07:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:31 crc kubenswrapper[4953]: I1203 00:07:31.503332 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-wj6rz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"96aa67a0-15dd-420d-a1d5-ed239d7df1fb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01815cfd3f58c3f3f5f2a894c1790e0b69c5060c48dc0b1deff050b6ad7af916\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kf62m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b12bfb5bdf9fd4f2c50a335fc82df6ee0440eaf8741897da03aa2dae368cc6b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b12bfb5bdf9fd4f2c50a335fc82df6ee0440eaf8741897da03aa2dae368cc6b4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:07:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kf62m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83aaa5fa28110bc0ef05aff75cd1442bc9bbdd33b822ae2ef22d4e4ab83238bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://83aaa5fa28110bc0ef05aff75cd1442bc9bbdd33b822ae2ef22d4e4ab83238bc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:07:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:07:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kf62m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ab6213a33a19fedebce6d74a0ebbb3b3a2de5687504e21cc24093cc42ade635\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ab6213a33a19fedebce6d74a0ebbb3b3a2de5687504e21cc24093cc42ade635\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:07:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:07:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kf62m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://de480488843afa6b8dede1e1b42b5181dca1993e2f39dc1b461a18a57e3668e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://de480488843afa6b8dede1e1b42b5181dca1993e2f39dc1b461a18a57e3668e1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:07:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kf62m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://97be4ba692bc61078d4b0ba39fb0c12772548c6c9251ff3356ea702f151223cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://97be4ba692bc61078d4b0ba39fb0c12772548c6c9251ff3356ea702f151223cd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:07:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kf62m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1a1f009fcbd8af9217db1dc97f662719f8c1db232df0e2804e44c881abb1df2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b1a1f009fcbd8af9217db1dc97f662719f8c1db232df0e2804e44c881abb1df2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:07:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:07:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kf62m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:07:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-wj6rz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:31Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:31 crc kubenswrapper[4953]: I1203 00:07:31.527305 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-txffx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8816ebee-c735-4b8d-b1ec-65349707b347\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:16Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:16Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4b85224419a4075acb340e4b11095f2adb5c316e285afa4e19295c2181c0d24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0f6f645a58993f052a6f365bb808f62d99f7fea74886c5cc022b2d86112321a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a61b951a9cbab2c48c0f2df8351f97d7ff71bf3674f641fe5eac30155232b224\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd5f428de229749485267e81d34e31b708728b52474032e88460e54747f21e11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://00a2ea4a9bb870fce30e7fa845f787f184e7a19e42c95511655ea8c5f84a26d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc0a5d4cc9d614d8492626a1b114bdc52f510eb5965d842c02c083f08d4849a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://efbfc74a8598c47a658c06b597fa36d2a5ac8d02265d2864fd2832dd11ecb2cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aecc0ae6dd34893c9023da3b94ca01ae96ac1e7cd7f5fd18748cce6c273a7f9e\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T00:07:26Z\\\",\\\"message\\\":\\\"ller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1203 00:07:25.992760 6257 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1203 00:07:25.992789 6257 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1203 00:07:25.992837 6257 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1203 00:07:25.992889 6257 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1203 00:07:25.992927 6257 handler.go:208] Removed *v1.Node event handler 7\\\\nI1203 00:07:25.992967 6257 handler.go:208] Removed *v1.Node event handler 2\\\\nI1203 00:07:25.992993 6257 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1203 00:07:25.992947 6257 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1203 00:07:25.993079 6257 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1203 00:07:25.992969 6257 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1203 00:07:25.993139 6257 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1203 00:07:25.993168 6257 factory.go:656] Stopping watch factory\\\\nI1203 00:07:25.993190 6257 ovnkube.go:599] Stopped ovnkube\\\\nI1203 00:07:25.993215 6257 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1203 00:07:25.993246 6257 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T00:07:23Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://efbfc74a8598c47a658c06b597fa36d2a5ac8d02265d2864fd2832dd11ecb2cc\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T00:07:29Z\\\",\\\"message\\\":\\\" 6400 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1203 00:07:29.614806 6400 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1203 00:07:29.614821 6400 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1203 00:07:29.614822 6400 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 00:07:29.614854 6400 factory.go:656] Stopping watch factory\\\\nI1203 00:07:29.614869 6400 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1203 00:07:29.614878 6400 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1203 00:07:29.614875 6400 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1203 00:07:29.614894 6400 handler.go:208] Removed *v1.Node event handler 7\\\\nI1203 00:07:29.614906 6400 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1203 00:07:29.614915 6400 reflector.go:311] Stopping reflector *v1.NetworkAttachmentDefinition (0s) from github.com/k8snetworkplumbingwg/network-attachment-definition-client/pkg/client/informers/externalversions/factory.go:117\\\\nI1203 00:07:29.615144 6400 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 00:07:29.615318 6400 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T00:07:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7deadc8e135ff1bc7c5a2f093e489bb5b4b6f166bc874c6acc2a841c5397f11e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af4c7764fca87144d01f3738ea010308744ccf50c54b7c471ccc213112de00d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af4c7764fca87144d01f3738ea010308744ccf50c54b7c471ccc213112de00d5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:07:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:07:16Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-txffx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:31Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:31 crc kubenswrapper[4953]: I1203 00:07:31.539865 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5cmlc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7e11a1d3-ecee-4bf0-87bf-085fab97b73d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d42390b9c53efc63add8ddcebf03c976f5e5448eaa2c34e076b2dc90dbac60fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6x2sc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:07:18Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5cmlc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:31Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:31 crc kubenswrapper[4953]: I1203 00:07:31.553774 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-chmnz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a12fe6d-d25a-4b08-90da-cc9e510980a7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:29Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:29Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c982g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c982g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:07:29Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-chmnz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:31Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:31 crc kubenswrapper[4953]: I1203 00:07:31.565096 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-xshk4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bd03a758-0ca9-47d8-86f7-18b2d610df43\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47gfh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47gfh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:07:31Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-xshk4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:31Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:31 crc kubenswrapper[4953]: I1203 00:07:31.584417 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cf821840-5353-4620-8f73-99b56b371ab0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cae5e2994fcc24dc9b2ea73e79572615b3311edc42642b7f278fb84123989b23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://29e4b9d1b70a93fb133e0759ad148ae133766a66efeae3ff06a21dfae0e805ff\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dec43536772aa7c3f5a662fa7f3bd73b1cae2791ad454f85792287828d0ffeed\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://034e0bac6892d361700b8f33d6427b8a55e2bd3f89b14ea77a374c6be94d25e2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://761973393a93c59c23e69f41b8ac0d16abc35abd61ecbc6d51a42d94cd664233\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T00:07:09Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 00:07:03.713374 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 00:07:03.715052 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1528563048/tls.crt::/tmp/serving-cert-1528563048/tls.key\\\\\\\"\\\\nI1203 00:07:09.300052 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 00:07:09.304192 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 00:07:09.304255 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 00:07:09.304309 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 00:07:09.304323 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 00:07:09.313563 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 00:07:09.313597 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 00:07:09.313603 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 00:07:09.313609 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 00:07:09.313613 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 00:07:09.313618 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 00:07:09.313622 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 00:07:09.313709 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 00:07:09.327347 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T00:06:53Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c143d4232fb8bf7ef5999a664e431a00ab0db030f4d31a28d0bb21e00e38581\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:53Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72af6c27296eae2e275d455795c68f0d272869d63642b8353b17d466d0c901cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://72af6c27296eae2e275d455795c68f0d272869d63642b8353b17d466d0c901cf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:06:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:06:51Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:31Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:31 crc kubenswrapper[4953]: I1203 00:07:31.596914 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:31 crc kubenswrapper[4953]: I1203 00:07:31.596951 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:31 crc kubenswrapper[4953]: I1203 00:07:31.596962 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:31 crc kubenswrapper[4953]: I1203 00:07:31.596995 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:31 crc kubenswrapper[4953]: I1203 00:07:31.597009 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:31Z","lastTransitionTime":"2025-12-03T00:07:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:31 crc kubenswrapper[4953]: I1203 00:07:31.599616 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:10Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:31Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:31 crc kubenswrapper[4953]: I1203 00:07:31.616095 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce8b414e1fc22eb30decb6a6411b672e7f443055cb89ebb029fbf6ffc1d5dca7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:31Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:31 crc kubenswrapper[4953]: I1203 00:07:31.631669 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-6dfbn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53ffb1e9-9a2e-4869-80d7-776f4aa7ef8b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ab4205c0d38ad6a0c32077eadb1fd5fd3a279d9c8b8a2ffbe693ec2c4018370c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ct26c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7df84405eda520ad5a22d35750edf07ef3c3977a756ad5739c54f91cba9da7b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ct26c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:07:16Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-6dfbn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:31Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:31 crc kubenswrapper[4953]: I1203 00:07:31.647228 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:31Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:31 crc kubenswrapper[4953]: I1203 00:07:31.658210 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2ed9c7fa96c8951fbc1728fe7a91fefdd999077285621df9c1abedf9ae0c449\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:31Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:31 crc kubenswrapper[4953]: I1203 00:07:31.673711 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:10Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:31Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:31 crc kubenswrapper[4953]: I1203 00:07:31.681221 4953 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-txffx_8816ebee-c735-4b8d-b1ec-65349707b347/ovnkube-controller/1.log" Dec 03 00:07:31 crc kubenswrapper[4953]: I1203 00:07:31.686019 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-chmnz" event={"ID":"7a12fe6d-d25a-4b08-90da-cc9e510980a7","Type":"ContainerStarted","Data":"6f802d5628a8cc32618044b9383211d5a492c56060d427f953b7793a1064029e"} Dec 03 00:07:31 crc kubenswrapper[4953]: I1203 00:07:31.692532 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-675qk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"43a4bad8-71f4-441a-8302-aeb526a986af\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5b5794fef6477aab5deed8f642e8cc50b37eb1fd0cc6f19c003f34ad31c1e3c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gdrw4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:07:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-675qk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:31Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:31 crc kubenswrapper[4953]: I1203 00:07:31.698338 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:31 crc kubenswrapper[4953]: I1203 00:07:31.698386 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:31 crc kubenswrapper[4953]: I1203 00:07:31.698403 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:31 crc kubenswrapper[4953]: I1203 00:07:31.698426 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:31 crc kubenswrapper[4953]: I1203 00:07:31.698450 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:31Z","lastTransitionTime":"2025-12-03T00:07:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:31 crc kubenswrapper[4953]: I1203 00:07:31.706046 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-kl5s9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d85e5c97-16ad-4bd7-9e73-b7f12678be97\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c21df91a9e4e3165686882b2eb559498224063dc633c3a63f958598108eca29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j24lb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:07:16Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-kl5s9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:31Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:31 crc kubenswrapper[4953]: I1203 00:07:31.720107 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f94f8c51-c665-4ca3-bf17-7a04ea5c7ade\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72c5cd7dc1c3a8adfedab6ba43a59d80b3efe65f2bb56be95e76cc86a409dcf3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2810ffc1b141fa7da811406053db9d72c708edf5f2bb1dd75e4d52ab35fd24c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c67b4277b20a9f1287f50ef82d0785c9d02543a8e7ed26e0413eae332b866db7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://258d78e1ad9d3dcda6976909ae3fe71aecead5b995fc66b3fd6e2722d80b5ab8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:06:51Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:31Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:31 crc kubenswrapper[4953]: I1203 00:07:31.738851 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09657a9f448a4c92a759c73a0a892076d7eb46393e6fd565e5f73f34c71aac23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5af5eac4a58870f63497f99b503cfa5fb922278095e0332ec50376925ecf2109\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:31Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:31 crc kubenswrapper[4953]: I1203 00:07:31.759717 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-wj6rz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"96aa67a0-15dd-420d-a1d5-ed239d7df1fb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01815cfd3f58c3f3f5f2a894c1790e0b69c5060c48dc0b1deff050b6ad7af916\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kf62m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b12bfb5bdf9fd4f2c50a335fc82df6ee0440eaf8741897da03aa2dae368cc6b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b12bfb5bdf9fd4f2c50a335fc82df6ee0440eaf8741897da03aa2dae368cc6b4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:07:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kf62m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83aaa5fa28110bc0ef05aff75cd1442bc9bbdd33b822ae2ef22d4e4ab83238bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://83aaa5fa28110bc0ef05aff75cd1442bc9bbdd33b822ae2ef22d4e4ab83238bc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:07:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:07:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kf62m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ab6213a33a19fedebce6d74a0ebbb3b3a2de5687504e21cc24093cc42ade635\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ab6213a33a19fedebce6d74a0ebbb3b3a2de5687504e21cc24093cc42ade635\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:07:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:07:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kf62m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://de480488843afa6b8dede1e1b42b5181dca1993e2f39dc1b461a18a57e3668e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://de480488843afa6b8dede1e1b42b5181dca1993e2f39dc1b461a18a57e3668e1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:07:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kf62m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://97be4ba692bc61078d4b0ba39fb0c12772548c6c9251ff3356ea702f151223cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://97be4ba692bc61078d4b0ba39fb0c12772548c6c9251ff3356ea702f151223cd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:07:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kf62m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1a1f009fcbd8af9217db1dc97f662719f8c1db232df0e2804e44c881abb1df2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b1a1f009fcbd8af9217db1dc97f662719f8c1db232df0e2804e44c881abb1df2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:07:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:07:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kf62m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:07:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-wj6rz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:31Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:31 crc kubenswrapper[4953]: I1203 00:07:31.786365 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-txffx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8816ebee-c735-4b8d-b1ec-65349707b347\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:16Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:16Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4b85224419a4075acb340e4b11095f2adb5c316e285afa4e19295c2181c0d24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0f6f645a58993f052a6f365bb808f62d99f7fea74886c5cc022b2d86112321a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a61b951a9cbab2c48c0f2df8351f97d7ff71bf3674f641fe5eac30155232b224\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd5f428de229749485267e81d34e31b708728b52474032e88460e54747f21e11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://00a2ea4a9bb870fce30e7fa845f787f184e7a19e42c95511655ea8c5f84a26d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc0a5d4cc9d614d8492626a1b114bdc52f510eb5965d842c02c083f08d4849a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://efbfc74a8598c47a658c06b597fa36d2a5ac8d02265d2864fd2832dd11ecb2cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aecc0ae6dd34893c9023da3b94ca01ae96ac1e7cd7f5fd18748cce6c273a7f9e\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T00:07:26Z\\\",\\\"message\\\":\\\"ller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1203 00:07:25.992760 6257 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1203 00:07:25.992789 6257 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1203 00:07:25.992837 6257 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1203 00:07:25.992889 6257 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1203 00:07:25.992927 6257 handler.go:208] Removed *v1.Node event handler 7\\\\nI1203 00:07:25.992967 6257 handler.go:208] Removed *v1.Node event handler 2\\\\nI1203 00:07:25.992993 6257 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1203 00:07:25.992947 6257 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1203 00:07:25.993079 6257 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1203 00:07:25.992969 6257 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1203 00:07:25.993139 6257 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1203 00:07:25.993168 6257 factory.go:656] Stopping watch factory\\\\nI1203 00:07:25.993190 6257 ovnkube.go:599] Stopped ovnkube\\\\nI1203 00:07:25.993215 6257 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1203 00:07:25.993246 6257 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T00:07:23Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://efbfc74a8598c47a658c06b597fa36d2a5ac8d02265d2864fd2832dd11ecb2cc\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T00:07:29Z\\\",\\\"message\\\":\\\" 6400 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1203 00:07:29.614806 6400 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1203 00:07:29.614821 6400 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1203 00:07:29.614822 6400 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 00:07:29.614854 6400 factory.go:656] Stopping watch factory\\\\nI1203 00:07:29.614869 6400 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1203 00:07:29.614878 6400 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1203 00:07:29.614875 6400 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1203 00:07:29.614894 6400 handler.go:208] Removed *v1.Node event handler 7\\\\nI1203 00:07:29.614906 6400 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1203 00:07:29.614915 6400 reflector.go:311] Stopping reflector *v1.NetworkAttachmentDefinition (0s) from github.com/k8snetworkplumbingwg/network-attachment-definition-client/pkg/client/informers/externalversions/factory.go:117\\\\nI1203 00:07:29.615144 6400 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 00:07:29.615318 6400 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T00:07:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7deadc8e135ff1bc7c5a2f093e489bb5b4b6f166bc874c6acc2a841c5397f11e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af4c7764fca87144d01f3738ea010308744ccf50c54b7c471ccc213112de00d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af4c7764fca87144d01f3738ea010308744ccf50c54b7c471ccc213112de00d5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:07:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:07:16Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-txffx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:31Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:31 crc kubenswrapper[4953]: I1203 00:07:31.790788 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/bd03a758-0ca9-47d8-86f7-18b2d610df43-metrics-certs\") pod \"network-metrics-daemon-xshk4\" (UID: \"bd03a758-0ca9-47d8-86f7-18b2d610df43\") " pod="openshift-multus/network-metrics-daemon-xshk4" Dec 03 00:07:31 crc kubenswrapper[4953]: E1203 00:07:31.791018 4953 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 03 00:07:31 crc kubenswrapper[4953]: E1203 00:07:31.791090 4953 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/bd03a758-0ca9-47d8-86f7-18b2d610df43-metrics-certs podName:bd03a758-0ca9-47d8-86f7-18b2d610df43 nodeName:}" failed. No retries permitted until 2025-12-03 00:07:32.791070144 +0000 UTC m=+41.775350937 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/bd03a758-0ca9-47d8-86f7-18b2d610df43-metrics-certs") pod "network-metrics-daemon-xshk4" (UID: "bd03a758-0ca9-47d8-86f7-18b2d610df43") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 03 00:07:31 crc kubenswrapper[4953]: I1203 00:07:31.798191 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5cmlc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7e11a1d3-ecee-4bf0-87bf-085fab97b73d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d42390b9c53efc63add8ddcebf03c976f5e5448eaa2c34e076b2dc90dbac60fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6x2sc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:07:18Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5cmlc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:31Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:31 crc kubenswrapper[4953]: I1203 00:07:31.800477 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:31 crc kubenswrapper[4953]: I1203 00:07:31.800515 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:31 crc kubenswrapper[4953]: I1203 00:07:31.800530 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:31 crc kubenswrapper[4953]: I1203 00:07:31.800550 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:31 crc kubenswrapper[4953]: I1203 00:07:31.800565 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:31Z","lastTransitionTime":"2025-12-03T00:07:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:31 crc kubenswrapper[4953]: I1203 00:07:31.811750 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-chmnz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a12fe6d-d25a-4b08-90da-cc9e510980a7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://711ab839cb9c0a90554dadbd45d810fc78aac09739340482fef8de8743b77283\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c982g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f802d5628a8cc32618044b9383211d5a492c56060d427f953b7793a1064029e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c982g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:07:29Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-chmnz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:31Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:31 crc kubenswrapper[4953]: I1203 00:07:31.823443 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-xshk4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bd03a758-0ca9-47d8-86f7-18b2d610df43\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47gfh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47gfh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:07:31Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-xshk4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:31Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:31 crc kubenswrapper[4953]: I1203 00:07:31.845424 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cf821840-5353-4620-8f73-99b56b371ab0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cae5e2994fcc24dc9b2ea73e79572615b3311edc42642b7f278fb84123989b23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://29e4b9d1b70a93fb133e0759ad148ae133766a66efeae3ff06a21dfae0e805ff\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dec43536772aa7c3f5a662fa7f3bd73b1cae2791ad454f85792287828d0ffeed\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://034e0bac6892d361700b8f33d6427b8a55e2bd3f89b14ea77a374c6be94d25e2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://761973393a93c59c23e69f41b8ac0d16abc35abd61ecbc6d51a42d94cd664233\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T00:07:09Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 00:07:03.713374 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 00:07:03.715052 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1528563048/tls.crt::/tmp/serving-cert-1528563048/tls.key\\\\\\\"\\\\nI1203 00:07:09.300052 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 00:07:09.304192 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 00:07:09.304255 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 00:07:09.304309 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 00:07:09.304323 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 00:07:09.313563 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 00:07:09.313597 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 00:07:09.313603 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 00:07:09.313609 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 00:07:09.313613 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 00:07:09.313618 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 00:07:09.313622 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 00:07:09.313709 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 00:07:09.327347 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T00:06:53Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c143d4232fb8bf7ef5999a664e431a00ab0db030f4d31a28d0bb21e00e38581\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:53Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72af6c27296eae2e275d455795c68f0d272869d63642b8353b17d466d0c901cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://72af6c27296eae2e275d455795c68f0d272869d63642b8353b17d466d0c901cf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:06:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:06:51Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:31Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:31 crc kubenswrapper[4953]: I1203 00:07:31.862329 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:10Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:31Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:31 crc kubenswrapper[4953]: I1203 00:07:31.876450 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce8b414e1fc22eb30decb6a6411b672e7f443055cb89ebb029fbf6ffc1d5dca7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:31Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:31 crc kubenswrapper[4953]: I1203 00:07:31.891763 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-6dfbn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53ffb1e9-9a2e-4869-80d7-776f4aa7ef8b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ab4205c0d38ad6a0c32077eadb1fd5fd3a279d9c8b8a2ffbe693ec2c4018370c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ct26c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7df84405eda520ad5a22d35750edf07ef3c3977a756ad5739c54f91cba9da7b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ct26c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:07:16Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-6dfbn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:31Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:31 crc kubenswrapper[4953]: I1203 00:07:31.902847 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:31 crc kubenswrapper[4953]: I1203 00:07:31.902884 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:31 crc kubenswrapper[4953]: I1203 00:07:31.902895 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:31 crc kubenswrapper[4953]: I1203 00:07:31.902910 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:31 crc kubenswrapper[4953]: I1203 00:07:31.902922 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:31Z","lastTransitionTime":"2025-12-03T00:07:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:31 crc kubenswrapper[4953]: I1203 00:07:31.906160 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:31Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:31 crc kubenswrapper[4953]: I1203 00:07:31.919998 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2ed9c7fa96c8951fbc1728fe7a91fefdd999077285621df9c1abedf9ae0c449\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:31Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:31 crc kubenswrapper[4953]: I1203 00:07:31.932672 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:10Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:31Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:31 crc kubenswrapper[4953]: I1203 00:07:31.950183 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-675qk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"43a4bad8-71f4-441a-8302-aeb526a986af\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5b5794fef6477aab5deed8f642e8cc50b37eb1fd0cc6f19c003f34ad31c1e3c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gdrw4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:07:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-675qk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:31Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:32 crc kubenswrapper[4953]: I1203 00:07:32.005749 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:32 crc kubenswrapper[4953]: I1203 00:07:32.005784 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:32 crc kubenswrapper[4953]: I1203 00:07:32.005794 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:32 crc kubenswrapper[4953]: I1203 00:07:32.005872 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:32 crc kubenswrapper[4953]: I1203 00:07:32.005885 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:32Z","lastTransitionTime":"2025-12-03T00:07:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:32 crc kubenswrapper[4953]: I1203 00:07:32.108921 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:32 crc kubenswrapper[4953]: I1203 00:07:32.108961 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:32 crc kubenswrapper[4953]: I1203 00:07:32.108996 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:32 crc kubenswrapper[4953]: I1203 00:07:32.109015 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:32 crc kubenswrapper[4953]: I1203 00:07:32.109026 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:32Z","lastTransitionTime":"2025-12-03T00:07:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:32 crc kubenswrapper[4953]: I1203 00:07:32.212461 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:32 crc kubenswrapper[4953]: I1203 00:07:32.212530 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:32 crc kubenswrapper[4953]: I1203 00:07:32.212555 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:32 crc kubenswrapper[4953]: I1203 00:07:32.212583 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:32 crc kubenswrapper[4953]: I1203 00:07:32.212602 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:32Z","lastTransitionTime":"2025-12-03T00:07:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:32 crc kubenswrapper[4953]: I1203 00:07:32.320733 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:32 crc kubenswrapper[4953]: I1203 00:07:32.320781 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:32 crc kubenswrapper[4953]: I1203 00:07:32.320799 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:32 crc kubenswrapper[4953]: I1203 00:07:32.320823 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:32 crc kubenswrapper[4953]: I1203 00:07:32.320840 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:32Z","lastTransitionTime":"2025-12-03T00:07:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:32 crc kubenswrapper[4953]: I1203 00:07:32.347170 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 00:07:32 crc kubenswrapper[4953]: I1203 00:07:32.347292 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 00:07:32 crc kubenswrapper[4953]: I1203 00:07:32.347175 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 00:07:32 crc kubenswrapper[4953]: E1203 00:07:32.347349 4953 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 00:07:32 crc kubenswrapper[4953]: E1203 00:07:32.347463 4953 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 00:07:32 crc kubenswrapper[4953]: E1203 00:07:32.347613 4953 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 00:07:32 crc kubenswrapper[4953]: I1203 00:07:32.424792 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:32 crc kubenswrapper[4953]: I1203 00:07:32.424863 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:32 crc kubenswrapper[4953]: I1203 00:07:32.424887 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:32 crc kubenswrapper[4953]: I1203 00:07:32.424917 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:32 crc kubenswrapper[4953]: I1203 00:07:32.424943 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:32Z","lastTransitionTime":"2025-12-03T00:07:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:32 crc kubenswrapper[4953]: I1203 00:07:32.528423 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:32 crc kubenswrapper[4953]: I1203 00:07:32.528482 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:32 crc kubenswrapper[4953]: I1203 00:07:32.528501 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:32 crc kubenswrapper[4953]: I1203 00:07:32.528523 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:32 crc kubenswrapper[4953]: I1203 00:07:32.528540 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:32Z","lastTransitionTime":"2025-12-03T00:07:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:32 crc kubenswrapper[4953]: I1203 00:07:32.631667 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:32 crc kubenswrapper[4953]: I1203 00:07:32.631753 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:32 crc kubenswrapper[4953]: I1203 00:07:32.631772 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:32 crc kubenswrapper[4953]: I1203 00:07:32.631798 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:32 crc kubenswrapper[4953]: I1203 00:07:32.631820 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:32Z","lastTransitionTime":"2025-12-03T00:07:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:32 crc kubenswrapper[4953]: I1203 00:07:32.736249 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:32 crc kubenswrapper[4953]: I1203 00:07:32.736307 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:32 crc kubenswrapper[4953]: I1203 00:07:32.736325 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:32 crc kubenswrapper[4953]: I1203 00:07:32.736360 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:32 crc kubenswrapper[4953]: I1203 00:07:32.736385 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:32Z","lastTransitionTime":"2025-12-03T00:07:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:32 crc kubenswrapper[4953]: I1203 00:07:32.801901 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/bd03a758-0ca9-47d8-86f7-18b2d610df43-metrics-certs\") pod \"network-metrics-daemon-xshk4\" (UID: \"bd03a758-0ca9-47d8-86f7-18b2d610df43\") " pod="openshift-multus/network-metrics-daemon-xshk4" Dec 03 00:07:32 crc kubenswrapper[4953]: E1203 00:07:32.802118 4953 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 03 00:07:32 crc kubenswrapper[4953]: E1203 00:07:32.802211 4953 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/bd03a758-0ca9-47d8-86f7-18b2d610df43-metrics-certs podName:bd03a758-0ca9-47d8-86f7-18b2d610df43 nodeName:}" failed. No retries permitted until 2025-12-03 00:07:34.802185963 +0000 UTC m=+43.786466786 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/bd03a758-0ca9-47d8-86f7-18b2d610df43-metrics-certs") pod "network-metrics-daemon-xshk4" (UID: "bd03a758-0ca9-47d8-86f7-18b2d610df43") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 03 00:07:32 crc kubenswrapper[4953]: I1203 00:07:32.839529 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:32 crc kubenswrapper[4953]: I1203 00:07:32.839582 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:32 crc kubenswrapper[4953]: I1203 00:07:32.839598 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:32 crc kubenswrapper[4953]: I1203 00:07:32.839621 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:32 crc kubenswrapper[4953]: I1203 00:07:32.839640 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:32Z","lastTransitionTime":"2025-12-03T00:07:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:32 crc kubenswrapper[4953]: I1203 00:07:32.943129 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:32 crc kubenswrapper[4953]: I1203 00:07:32.943442 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:32 crc kubenswrapper[4953]: I1203 00:07:32.943555 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:32 crc kubenswrapper[4953]: I1203 00:07:32.943707 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:32 crc kubenswrapper[4953]: I1203 00:07:32.943868 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:32Z","lastTransitionTime":"2025-12-03T00:07:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:33 crc kubenswrapper[4953]: I1203 00:07:33.047301 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:33 crc kubenswrapper[4953]: I1203 00:07:33.047378 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:33 crc kubenswrapper[4953]: I1203 00:07:33.047397 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:33 crc kubenswrapper[4953]: I1203 00:07:33.047420 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:33 crc kubenswrapper[4953]: I1203 00:07:33.047481 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:33Z","lastTransitionTime":"2025-12-03T00:07:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:33 crc kubenswrapper[4953]: I1203 00:07:33.150534 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:33 crc kubenswrapper[4953]: I1203 00:07:33.150611 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:33 crc kubenswrapper[4953]: I1203 00:07:33.150634 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:33 crc kubenswrapper[4953]: I1203 00:07:33.150658 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:33 crc kubenswrapper[4953]: I1203 00:07:33.150675 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:33Z","lastTransitionTime":"2025-12-03T00:07:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:33 crc kubenswrapper[4953]: I1203 00:07:33.245478 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:33 crc kubenswrapper[4953]: I1203 00:07:33.245541 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:33 crc kubenswrapper[4953]: I1203 00:07:33.245559 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:33 crc kubenswrapper[4953]: I1203 00:07:33.245583 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:33 crc kubenswrapper[4953]: I1203 00:07:33.245600 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:33Z","lastTransitionTime":"2025-12-03T00:07:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:33 crc kubenswrapper[4953]: E1203 00:07:33.266337 4953 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:07:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:33Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:07:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:33Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:07:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:33Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:07:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:33Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"eb4ae5a6-0f6f-49ef-a027-1cdee9d3c536\\\",\\\"systemUUID\\\":\\\"4c1f3d44-3d3d-425d-993f-471223206a89\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:33Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:33 crc kubenswrapper[4953]: I1203 00:07:33.271917 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:33 crc kubenswrapper[4953]: I1203 00:07:33.272040 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:33 crc kubenswrapper[4953]: I1203 00:07:33.272065 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:33 crc kubenswrapper[4953]: I1203 00:07:33.272088 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:33 crc kubenswrapper[4953]: I1203 00:07:33.272106 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:33Z","lastTransitionTime":"2025-12-03T00:07:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:33 crc kubenswrapper[4953]: E1203 00:07:33.289496 4953 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:07:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:33Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:07:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:33Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:07:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:33Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:07:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:33Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"eb4ae5a6-0f6f-49ef-a027-1cdee9d3c536\\\",\\\"systemUUID\\\":\\\"4c1f3d44-3d3d-425d-993f-471223206a89\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:33Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:33 crc kubenswrapper[4953]: I1203 00:07:33.295952 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:33 crc kubenswrapper[4953]: I1203 00:07:33.296011 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:33 crc kubenswrapper[4953]: I1203 00:07:33.296023 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:33 crc kubenswrapper[4953]: I1203 00:07:33.296039 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:33 crc kubenswrapper[4953]: I1203 00:07:33.296052 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:33Z","lastTransitionTime":"2025-12-03T00:07:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:33 crc kubenswrapper[4953]: E1203 00:07:33.314684 4953 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:07:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:33Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:07:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:33Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:07:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:33Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:07:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:33Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"eb4ae5a6-0f6f-49ef-a027-1cdee9d3c536\\\",\\\"systemUUID\\\":\\\"4c1f3d44-3d3d-425d-993f-471223206a89\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:33Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:33 crc kubenswrapper[4953]: I1203 00:07:33.319556 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:33 crc kubenswrapper[4953]: I1203 00:07:33.319658 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:33 crc kubenswrapper[4953]: I1203 00:07:33.319687 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:33 crc kubenswrapper[4953]: I1203 00:07:33.319722 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:33 crc kubenswrapper[4953]: I1203 00:07:33.319791 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:33Z","lastTransitionTime":"2025-12-03T00:07:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:33 crc kubenswrapper[4953]: E1203 00:07:33.341915 4953 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:07:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:33Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:07:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:33Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:07:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:33Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:07:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:33Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"eb4ae5a6-0f6f-49ef-a027-1cdee9d3c536\\\",\\\"systemUUID\\\":\\\"4c1f3d44-3d3d-425d-993f-471223206a89\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:33Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:33 crc kubenswrapper[4953]: I1203 00:07:33.347173 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xshk4" Dec 03 00:07:33 crc kubenswrapper[4953]: E1203 00:07:33.347417 4953 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xshk4" podUID="bd03a758-0ca9-47d8-86f7-18b2d610df43" Dec 03 00:07:33 crc kubenswrapper[4953]: I1203 00:07:33.348235 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:33 crc kubenswrapper[4953]: I1203 00:07:33.348298 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:33 crc kubenswrapper[4953]: I1203 00:07:33.348319 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:33 crc kubenswrapper[4953]: I1203 00:07:33.348344 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:33 crc kubenswrapper[4953]: I1203 00:07:33.348364 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:33Z","lastTransitionTime":"2025-12-03T00:07:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:33 crc kubenswrapper[4953]: E1203 00:07:33.370464 4953 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:07:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:33Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:07:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:33Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:07:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:33Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:07:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:33Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"eb4ae5a6-0f6f-49ef-a027-1cdee9d3c536\\\",\\\"systemUUID\\\":\\\"4c1f3d44-3d3d-425d-993f-471223206a89\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:33Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:33 crc kubenswrapper[4953]: E1203 00:07:33.370713 4953 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 03 00:07:33 crc kubenswrapper[4953]: I1203 00:07:33.372517 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:33 crc kubenswrapper[4953]: I1203 00:07:33.372554 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:33 crc kubenswrapper[4953]: I1203 00:07:33.372572 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:33 crc kubenswrapper[4953]: I1203 00:07:33.372592 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:33 crc kubenswrapper[4953]: I1203 00:07:33.372633 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:33Z","lastTransitionTime":"2025-12-03T00:07:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:33 crc kubenswrapper[4953]: I1203 00:07:33.475693 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:33 crc kubenswrapper[4953]: I1203 00:07:33.476311 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:33 crc kubenswrapper[4953]: I1203 00:07:33.476484 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:33 crc kubenswrapper[4953]: I1203 00:07:33.476629 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:33 crc kubenswrapper[4953]: I1203 00:07:33.476774 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:33Z","lastTransitionTime":"2025-12-03T00:07:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:33 crc kubenswrapper[4953]: I1203 00:07:33.580221 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:33 crc kubenswrapper[4953]: I1203 00:07:33.580283 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:33 crc kubenswrapper[4953]: I1203 00:07:33.580301 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:33 crc kubenswrapper[4953]: I1203 00:07:33.580326 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:33 crc kubenswrapper[4953]: I1203 00:07:33.580345 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:33Z","lastTransitionTime":"2025-12-03T00:07:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:33 crc kubenswrapper[4953]: I1203 00:07:33.684087 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:33 crc kubenswrapper[4953]: I1203 00:07:33.684472 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:33 crc kubenswrapper[4953]: I1203 00:07:33.684633 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:33 crc kubenswrapper[4953]: I1203 00:07:33.684801 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:33 crc kubenswrapper[4953]: I1203 00:07:33.684948 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:33Z","lastTransitionTime":"2025-12-03T00:07:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:33 crc kubenswrapper[4953]: I1203 00:07:33.787799 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:33 crc kubenswrapper[4953]: I1203 00:07:33.787874 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:33 crc kubenswrapper[4953]: I1203 00:07:33.787893 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:33 crc kubenswrapper[4953]: I1203 00:07:33.787918 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:33 crc kubenswrapper[4953]: I1203 00:07:33.787935 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:33Z","lastTransitionTime":"2025-12-03T00:07:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:33 crc kubenswrapper[4953]: I1203 00:07:33.890544 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:33 crc kubenswrapper[4953]: I1203 00:07:33.890618 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:33 crc kubenswrapper[4953]: I1203 00:07:33.890642 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:33 crc kubenswrapper[4953]: I1203 00:07:33.890673 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:33 crc kubenswrapper[4953]: I1203 00:07:33.890693 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:33Z","lastTransitionTime":"2025-12-03T00:07:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:33 crc kubenswrapper[4953]: I1203 00:07:33.993520 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:33 crc kubenswrapper[4953]: I1203 00:07:33.993675 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:33 crc kubenswrapper[4953]: I1203 00:07:33.993706 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:33 crc kubenswrapper[4953]: I1203 00:07:33.993730 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:33 crc kubenswrapper[4953]: I1203 00:07:33.993748 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:33Z","lastTransitionTime":"2025-12-03T00:07:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:34 crc kubenswrapper[4953]: I1203 00:07:34.097051 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:34 crc kubenswrapper[4953]: I1203 00:07:34.098029 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:34 crc kubenswrapper[4953]: I1203 00:07:34.098203 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:34 crc kubenswrapper[4953]: I1203 00:07:34.098338 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:34 crc kubenswrapper[4953]: I1203 00:07:34.098458 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:34Z","lastTransitionTime":"2025-12-03T00:07:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:34 crc kubenswrapper[4953]: I1203 00:07:34.200995 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:34 crc kubenswrapper[4953]: I1203 00:07:34.201035 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:34 crc kubenswrapper[4953]: I1203 00:07:34.201046 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:34 crc kubenswrapper[4953]: I1203 00:07:34.201059 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:34 crc kubenswrapper[4953]: I1203 00:07:34.201068 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:34Z","lastTransitionTime":"2025-12-03T00:07:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:34 crc kubenswrapper[4953]: I1203 00:07:34.303759 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:34 crc kubenswrapper[4953]: I1203 00:07:34.304288 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:34 crc kubenswrapper[4953]: I1203 00:07:34.304478 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:34 crc kubenswrapper[4953]: I1203 00:07:34.304634 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:34 crc kubenswrapper[4953]: I1203 00:07:34.304796 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:34Z","lastTransitionTime":"2025-12-03T00:07:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:34 crc kubenswrapper[4953]: I1203 00:07:34.347645 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 00:07:34 crc kubenswrapper[4953]: I1203 00:07:34.347706 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 00:07:34 crc kubenswrapper[4953]: E1203 00:07:34.347756 4953 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 00:07:34 crc kubenswrapper[4953]: E1203 00:07:34.347856 4953 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 00:07:34 crc kubenswrapper[4953]: I1203 00:07:34.348419 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 00:07:34 crc kubenswrapper[4953]: E1203 00:07:34.348765 4953 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 00:07:34 crc kubenswrapper[4953]: I1203 00:07:34.407776 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:34 crc kubenswrapper[4953]: I1203 00:07:34.408112 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:34 crc kubenswrapper[4953]: I1203 00:07:34.408289 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:34 crc kubenswrapper[4953]: I1203 00:07:34.408452 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:34 crc kubenswrapper[4953]: I1203 00:07:34.408676 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:34Z","lastTransitionTime":"2025-12-03T00:07:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:34 crc kubenswrapper[4953]: I1203 00:07:34.511550 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:34 crc kubenswrapper[4953]: I1203 00:07:34.511605 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:34 crc kubenswrapper[4953]: I1203 00:07:34.511621 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:34 crc kubenswrapper[4953]: I1203 00:07:34.511643 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:34 crc kubenswrapper[4953]: I1203 00:07:34.511661 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:34Z","lastTransitionTime":"2025-12-03T00:07:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:34 crc kubenswrapper[4953]: I1203 00:07:34.615022 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:34 crc kubenswrapper[4953]: I1203 00:07:34.615323 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:34 crc kubenswrapper[4953]: I1203 00:07:34.615404 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:34 crc kubenswrapper[4953]: I1203 00:07:34.615485 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:34 crc kubenswrapper[4953]: I1203 00:07:34.615553 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:34Z","lastTransitionTime":"2025-12-03T00:07:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:34 crc kubenswrapper[4953]: I1203 00:07:34.717624 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:34 crc kubenswrapper[4953]: I1203 00:07:34.717913 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:34 crc kubenswrapper[4953]: I1203 00:07:34.718048 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:34 crc kubenswrapper[4953]: I1203 00:07:34.718172 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:34 crc kubenswrapper[4953]: I1203 00:07:34.718260 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:34Z","lastTransitionTime":"2025-12-03T00:07:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:34 crc kubenswrapper[4953]: I1203 00:07:34.821717 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:34 crc kubenswrapper[4953]: I1203 00:07:34.822032 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:34 crc kubenswrapper[4953]: I1203 00:07:34.822134 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:34 crc kubenswrapper[4953]: I1203 00:07:34.822278 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:34 crc kubenswrapper[4953]: I1203 00:07:34.822356 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:34Z","lastTransitionTime":"2025-12-03T00:07:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:34 crc kubenswrapper[4953]: I1203 00:07:34.822940 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/bd03a758-0ca9-47d8-86f7-18b2d610df43-metrics-certs\") pod \"network-metrics-daemon-xshk4\" (UID: \"bd03a758-0ca9-47d8-86f7-18b2d610df43\") " pod="openshift-multus/network-metrics-daemon-xshk4" Dec 03 00:07:34 crc kubenswrapper[4953]: E1203 00:07:34.823152 4953 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 03 00:07:34 crc kubenswrapper[4953]: E1203 00:07:34.823266 4953 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/bd03a758-0ca9-47d8-86f7-18b2d610df43-metrics-certs podName:bd03a758-0ca9-47d8-86f7-18b2d610df43 nodeName:}" failed. No retries permitted until 2025-12-03 00:07:38.82323782 +0000 UTC m=+47.807518633 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/bd03a758-0ca9-47d8-86f7-18b2d610df43-metrics-certs") pod "network-metrics-daemon-xshk4" (UID: "bd03a758-0ca9-47d8-86f7-18b2d610df43") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 03 00:07:34 crc kubenswrapper[4953]: I1203 00:07:34.925406 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:34 crc kubenswrapper[4953]: I1203 00:07:34.925483 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:34 crc kubenswrapper[4953]: I1203 00:07:34.925504 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:34 crc kubenswrapper[4953]: I1203 00:07:34.925567 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:34 crc kubenswrapper[4953]: I1203 00:07:34.925591 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:34Z","lastTransitionTime":"2025-12-03T00:07:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:35 crc kubenswrapper[4953]: I1203 00:07:35.027810 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:35 crc kubenswrapper[4953]: I1203 00:07:35.028141 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:35 crc kubenswrapper[4953]: I1203 00:07:35.028332 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:35 crc kubenswrapper[4953]: I1203 00:07:35.028553 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:35 crc kubenswrapper[4953]: I1203 00:07:35.028761 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:35Z","lastTransitionTime":"2025-12-03T00:07:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:35 crc kubenswrapper[4953]: I1203 00:07:35.131450 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:35 crc kubenswrapper[4953]: I1203 00:07:35.132009 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:35 crc kubenswrapper[4953]: I1203 00:07:35.132098 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:35 crc kubenswrapper[4953]: I1203 00:07:35.132177 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:35 crc kubenswrapper[4953]: I1203 00:07:35.132262 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:35Z","lastTransitionTime":"2025-12-03T00:07:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:35 crc kubenswrapper[4953]: I1203 00:07:35.234765 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:35 crc kubenswrapper[4953]: I1203 00:07:35.235145 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:35 crc kubenswrapper[4953]: I1203 00:07:35.235267 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:35 crc kubenswrapper[4953]: I1203 00:07:35.235361 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:35 crc kubenswrapper[4953]: I1203 00:07:35.235449 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:35Z","lastTransitionTime":"2025-12-03T00:07:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:35 crc kubenswrapper[4953]: I1203 00:07:35.338758 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:35 crc kubenswrapper[4953]: I1203 00:07:35.339055 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:35 crc kubenswrapper[4953]: I1203 00:07:35.339153 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:35 crc kubenswrapper[4953]: I1203 00:07:35.339237 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:35 crc kubenswrapper[4953]: I1203 00:07:35.339320 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:35Z","lastTransitionTime":"2025-12-03T00:07:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:35 crc kubenswrapper[4953]: I1203 00:07:35.348327 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xshk4" Dec 03 00:07:35 crc kubenswrapper[4953]: E1203 00:07:35.348462 4953 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xshk4" podUID="bd03a758-0ca9-47d8-86f7-18b2d610df43" Dec 03 00:07:35 crc kubenswrapper[4953]: I1203 00:07:35.441540 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:35 crc kubenswrapper[4953]: I1203 00:07:35.441591 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:35 crc kubenswrapper[4953]: I1203 00:07:35.441606 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:35 crc kubenswrapper[4953]: I1203 00:07:35.441626 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:35 crc kubenswrapper[4953]: I1203 00:07:35.441641 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:35Z","lastTransitionTime":"2025-12-03T00:07:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:35 crc kubenswrapper[4953]: I1203 00:07:35.544965 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:35 crc kubenswrapper[4953]: I1203 00:07:35.545024 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:35 crc kubenswrapper[4953]: I1203 00:07:35.545036 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:35 crc kubenswrapper[4953]: I1203 00:07:35.545055 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:35 crc kubenswrapper[4953]: I1203 00:07:35.545068 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:35Z","lastTransitionTime":"2025-12-03T00:07:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:35 crc kubenswrapper[4953]: I1203 00:07:35.647518 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:35 crc kubenswrapper[4953]: I1203 00:07:35.647575 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:35 crc kubenswrapper[4953]: I1203 00:07:35.647590 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:35 crc kubenswrapper[4953]: I1203 00:07:35.647612 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:35 crc kubenswrapper[4953]: I1203 00:07:35.647627 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:35Z","lastTransitionTime":"2025-12-03T00:07:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:35 crc kubenswrapper[4953]: I1203 00:07:35.750372 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:35 crc kubenswrapper[4953]: I1203 00:07:35.750780 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:35 crc kubenswrapper[4953]: I1203 00:07:35.750943 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:35 crc kubenswrapper[4953]: I1203 00:07:35.751216 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:35 crc kubenswrapper[4953]: I1203 00:07:35.751261 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:35Z","lastTransitionTime":"2025-12-03T00:07:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:35 crc kubenswrapper[4953]: I1203 00:07:35.854266 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:35 crc kubenswrapper[4953]: I1203 00:07:35.854666 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:35 crc kubenswrapper[4953]: I1203 00:07:35.854842 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:35 crc kubenswrapper[4953]: I1203 00:07:35.855158 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:35 crc kubenswrapper[4953]: I1203 00:07:35.855381 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:35Z","lastTransitionTime":"2025-12-03T00:07:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:35 crc kubenswrapper[4953]: I1203 00:07:35.958056 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:35 crc kubenswrapper[4953]: I1203 00:07:35.958424 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:35 crc kubenswrapper[4953]: I1203 00:07:35.958640 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:35 crc kubenswrapper[4953]: I1203 00:07:35.958810 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:35 crc kubenswrapper[4953]: I1203 00:07:35.958957 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:35Z","lastTransitionTime":"2025-12-03T00:07:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:36 crc kubenswrapper[4953]: I1203 00:07:36.061776 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:36 crc kubenswrapper[4953]: I1203 00:07:36.062204 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:36 crc kubenswrapper[4953]: I1203 00:07:36.062416 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:36 crc kubenswrapper[4953]: I1203 00:07:36.062612 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:36 crc kubenswrapper[4953]: I1203 00:07:36.062850 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:36Z","lastTransitionTime":"2025-12-03T00:07:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:36 crc kubenswrapper[4953]: I1203 00:07:36.165835 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:36 crc kubenswrapper[4953]: I1203 00:07:36.165921 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:36 crc kubenswrapper[4953]: I1203 00:07:36.165955 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:36 crc kubenswrapper[4953]: I1203 00:07:36.166017 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:36 crc kubenswrapper[4953]: I1203 00:07:36.166036 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:36Z","lastTransitionTime":"2025-12-03T00:07:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:36 crc kubenswrapper[4953]: I1203 00:07:36.268942 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:36 crc kubenswrapper[4953]: I1203 00:07:36.269006 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:36 crc kubenswrapper[4953]: I1203 00:07:36.269016 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:36 crc kubenswrapper[4953]: I1203 00:07:36.269036 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:36 crc kubenswrapper[4953]: I1203 00:07:36.269049 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:36Z","lastTransitionTime":"2025-12-03T00:07:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:36 crc kubenswrapper[4953]: I1203 00:07:36.347740 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 00:07:36 crc kubenswrapper[4953]: I1203 00:07:36.347771 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 00:07:36 crc kubenswrapper[4953]: E1203 00:07:36.347999 4953 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 00:07:36 crc kubenswrapper[4953]: I1203 00:07:36.347776 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 00:07:36 crc kubenswrapper[4953]: E1203 00:07:36.348199 4953 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 00:07:36 crc kubenswrapper[4953]: E1203 00:07:36.348296 4953 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 00:07:36 crc kubenswrapper[4953]: I1203 00:07:36.372454 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:36 crc kubenswrapper[4953]: I1203 00:07:36.372516 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:36 crc kubenswrapper[4953]: I1203 00:07:36.372532 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:36 crc kubenswrapper[4953]: I1203 00:07:36.372557 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:36 crc kubenswrapper[4953]: I1203 00:07:36.372574 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:36Z","lastTransitionTime":"2025-12-03T00:07:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:36 crc kubenswrapper[4953]: I1203 00:07:36.475210 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:36 crc kubenswrapper[4953]: I1203 00:07:36.475274 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:36 crc kubenswrapper[4953]: I1203 00:07:36.475286 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:36 crc kubenswrapper[4953]: I1203 00:07:36.475303 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:36 crc kubenswrapper[4953]: I1203 00:07:36.475315 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:36Z","lastTransitionTime":"2025-12-03T00:07:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:36 crc kubenswrapper[4953]: I1203 00:07:36.578084 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:36 crc kubenswrapper[4953]: I1203 00:07:36.578125 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:36 crc kubenswrapper[4953]: I1203 00:07:36.578154 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:36 crc kubenswrapper[4953]: I1203 00:07:36.578173 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:36 crc kubenswrapper[4953]: I1203 00:07:36.578191 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:36Z","lastTransitionTime":"2025-12-03T00:07:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:36 crc kubenswrapper[4953]: I1203 00:07:36.681390 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:36 crc kubenswrapper[4953]: I1203 00:07:36.681458 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:36 crc kubenswrapper[4953]: I1203 00:07:36.681484 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:36 crc kubenswrapper[4953]: I1203 00:07:36.681519 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:36 crc kubenswrapper[4953]: I1203 00:07:36.681544 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:36Z","lastTransitionTime":"2025-12-03T00:07:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:36 crc kubenswrapper[4953]: I1203 00:07:36.785507 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:36 crc kubenswrapper[4953]: I1203 00:07:36.785575 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:36 crc kubenswrapper[4953]: I1203 00:07:36.785594 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:36 crc kubenswrapper[4953]: I1203 00:07:36.785620 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:36 crc kubenswrapper[4953]: I1203 00:07:36.785641 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:36Z","lastTransitionTime":"2025-12-03T00:07:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:36 crc kubenswrapper[4953]: I1203 00:07:36.888643 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:36 crc kubenswrapper[4953]: I1203 00:07:36.888709 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:36 crc kubenswrapper[4953]: I1203 00:07:36.888726 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:36 crc kubenswrapper[4953]: I1203 00:07:36.888751 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:36 crc kubenswrapper[4953]: I1203 00:07:36.888769 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:36Z","lastTransitionTime":"2025-12-03T00:07:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:36 crc kubenswrapper[4953]: I1203 00:07:36.992392 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:36 crc kubenswrapper[4953]: I1203 00:07:36.992449 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:36 crc kubenswrapper[4953]: I1203 00:07:36.992465 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:36 crc kubenswrapper[4953]: I1203 00:07:36.992488 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:36 crc kubenswrapper[4953]: I1203 00:07:36.992506 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:36Z","lastTransitionTime":"2025-12-03T00:07:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:37 crc kubenswrapper[4953]: I1203 00:07:37.095416 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:37 crc kubenswrapper[4953]: I1203 00:07:37.095481 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:37 crc kubenswrapper[4953]: I1203 00:07:37.095501 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:37 crc kubenswrapper[4953]: I1203 00:07:37.095527 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:37 crc kubenswrapper[4953]: I1203 00:07:37.095544 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:37Z","lastTransitionTime":"2025-12-03T00:07:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:37 crc kubenswrapper[4953]: I1203 00:07:37.198587 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:37 crc kubenswrapper[4953]: I1203 00:07:37.198643 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:37 crc kubenswrapper[4953]: I1203 00:07:37.198653 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:37 crc kubenswrapper[4953]: I1203 00:07:37.198673 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:37 crc kubenswrapper[4953]: I1203 00:07:37.198686 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:37Z","lastTransitionTime":"2025-12-03T00:07:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:37 crc kubenswrapper[4953]: I1203 00:07:37.301781 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:37 crc kubenswrapper[4953]: I1203 00:07:37.301840 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:37 crc kubenswrapper[4953]: I1203 00:07:37.301854 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:37 crc kubenswrapper[4953]: I1203 00:07:37.301876 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:37 crc kubenswrapper[4953]: I1203 00:07:37.301890 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:37Z","lastTransitionTime":"2025-12-03T00:07:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:37 crc kubenswrapper[4953]: I1203 00:07:37.347861 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xshk4" Dec 03 00:07:37 crc kubenswrapper[4953]: E1203 00:07:37.348156 4953 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xshk4" podUID="bd03a758-0ca9-47d8-86f7-18b2d610df43" Dec 03 00:07:37 crc kubenswrapper[4953]: I1203 00:07:37.405391 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:37 crc kubenswrapper[4953]: I1203 00:07:37.405437 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:37 crc kubenswrapper[4953]: I1203 00:07:37.405447 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:37 crc kubenswrapper[4953]: I1203 00:07:37.405463 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:37 crc kubenswrapper[4953]: I1203 00:07:37.405475 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:37Z","lastTransitionTime":"2025-12-03T00:07:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:37 crc kubenswrapper[4953]: I1203 00:07:37.508801 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:37 crc kubenswrapper[4953]: I1203 00:07:37.509139 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:37 crc kubenswrapper[4953]: I1203 00:07:37.509314 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:37 crc kubenswrapper[4953]: I1203 00:07:37.509471 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:37 crc kubenswrapper[4953]: I1203 00:07:37.509600 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:37Z","lastTransitionTime":"2025-12-03T00:07:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:37 crc kubenswrapper[4953]: I1203 00:07:37.613081 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:37 crc kubenswrapper[4953]: I1203 00:07:37.613477 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:37 crc kubenswrapper[4953]: I1203 00:07:37.613610 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:37 crc kubenswrapper[4953]: I1203 00:07:37.613771 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:37 crc kubenswrapper[4953]: I1203 00:07:37.613897 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:37Z","lastTransitionTime":"2025-12-03T00:07:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:37 crc kubenswrapper[4953]: I1203 00:07:37.716384 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:37 crc kubenswrapper[4953]: I1203 00:07:37.716481 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:37 crc kubenswrapper[4953]: I1203 00:07:37.716508 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:37 crc kubenswrapper[4953]: I1203 00:07:37.716540 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:37 crc kubenswrapper[4953]: I1203 00:07:37.716566 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:37Z","lastTransitionTime":"2025-12-03T00:07:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:37 crc kubenswrapper[4953]: I1203 00:07:37.820184 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:37 crc kubenswrapper[4953]: I1203 00:07:37.820244 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:37 crc kubenswrapper[4953]: I1203 00:07:37.820263 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:37 crc kubenswrapper[4953]: I1203 00:07:37.820289 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:37 crc kubenswrapper[4953]: I1203 00:07:37.820311 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:37Z","lastTransitionTime":"2025-12-03T00:07:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:37 crc kubenswrapper[4953]: I1203 00:07:37.923014 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:37 crc kubenswrapper[4953]: I1203 00:07:37.923058 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:37 crc kubenswrapper[4953]: I1203 00:07:37.923070 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:37 crc kubenswrapper[4953]: I1203 00:07:37.923088 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:37 crc kubenswrapper[4953]: I1203 00:07:37.923102 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:37Z","lastTransitionTime":"2025-12-03T00:07:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:38 crc kubenswrapper[4953]: I1203 00:07:38.026327 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:38 crc kubenswrapper[4953]: I1203 00:07:38.026408 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:38 crc kubenswrapper[4953]: I1203 00:07:38.026434 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:38 crc kubenswrapper[4953]: I1203 00:07:38.026469 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:38 crc kubenswrapper[4953]: I1203 00:07:38.026491 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:38Z","lastTransitionTime":"2025-12-03T00:07:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:38 crc kubenswrapper[4953]: I1203 00:07:38.130201 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:38 crc kubenswrapper[4953]: I1203 00:07:38.130262 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:38 crc kubenswrapper[4953]: I1203 00:07:38.130286 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:38 crc kubenswrapper[4953]: I1203 00:07:38.130315 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:38 crc kubenswrapper[4953]: I1203 00:07:38.130333 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:38Z","lastTransitionTime":"2025-12-03T00:07:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:38 crc kubenswrapper[4953]: I1203 00:07:38.234158 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:38 crc kubenswrapper[4953]: I1203 00:07:38.234225 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:38 crc kubenswrapper[4953]: I1203 00:07:38.234247 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:38 crc kubenswrapper[4953]: I1203 00:07:38.234275 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:38 crc kubenswrapper[4953]: I1203 00:07:38.234295 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:38Z","lastTransitionTime":"2025-12-03T00:07:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:38 crc kubenswrapper[4953]: I1203 00:07:38.337354 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:38 crc kubenswrapper[4953]: I1203 00:07:38.337401 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:38 crc kubenswrapper[4953]: I1203 00:07:38.337412 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:38 crc kubenswrapper[4953]: I1203 00:07:38.337430 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:38 crc kubenswrapper[4953]: I1203 00:07:38.337441 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:38Z","lastTransitionTime":"2025-12-03T00:07:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:38 crc kubenswrapper[4953]: I1203 00:07:38.347626 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 00:07:38 crc kubenswrapper[4953]: I1203 00:07:38.347727 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 00:07:38 crc kubenswrapper[4953]: I1203 00:07:38.347745 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 00:07:38 crc kubenswrapper[4953]: E1203 00:07:38.347866 4953 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 00:07:38 crc kubenswrapper[4953]: E1203 00:07:38.348007 4953 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 00:07:38 crc kubenswrapper[4953]: E1203 00:07:38.348153 4953 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 00:07:38 crc kubenswrapper[4953]: I1203 00:07:38.441204 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:38 crc kubenswrapper[4953]: I1203 00:07:38.441275 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:38 crc kubenswrapper[4953]: I1203 00:07:38.441301 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:38 crc kubenswrapper[4953]: I1203 00:07:38.441330 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:38 crc kubenswrapper[4953]: I1203 00:07:38.441353 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:38Z","lastTransitionTime":"2025-12-03T00:07:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:38 crc kubenswrapper[4953]: I1203 00:07:38.544852 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:38 crc kubenswrapper[4953]: I1203 00:07:38.544912 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:38 crc kubenswrapper[4953]: I1203 00:07:38.544931 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:38 crc kubenswrapper[4953]: I1203 00:07:38.544956 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:38 crc kubenswrapper[4953]: I1203 00:07:38.545016 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:38Z","lastTransitionTime":"2025-12-03T00:07:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:38 crc kubenswrapper[4953]: I1203 00:07:38.648633 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:38 crc kubenswrapper[4953]: I1203 00:07:38.648725 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:38 crc kubenswrapper[4953]: I1203 00:07:38.648737 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:38 crc kubenswrapper[4953]: I1203 00:07:38.648758 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:38 crc kubenswrapper[4953]: I1203 00:07:38.648771 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:38Z","lastTransitionTime":"2025-12-03T00:07:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:38 crc kubenswrapper[4953]: I1203 00:07:38.752162 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:38 crc kubenswrapper[4953]: I1203 00:07:38.752248 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:38 crc kubenswrapper[4953]: I1203 00:07:38.752269 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:38 crc kubenswrapper[4953]: I1203 00:07:38.752302 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:38 crc kubenswrapper[4953]: I1203 00:07:38.752326 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:38Z","lastTransitionTime":"2025-12-03T00:07:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:38 crc kubenswrapper[4953]: I1203 00:07:38.856507 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:38 crc kubenswrapper[4953]: I1203 00:07:38.856941 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:38 crc kubenswrapper[4953]: I1203 00:07:38.857001 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:38 crc kubenswrapper[4953]: I1203 00:07:38.857031 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:38 crc kubenswrapper[4953]: I1203 00:07:38.857049 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:38Z","lastTransitionTime":"2025-12-03T00:07:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:38 crc kubenswrapper[4953]: I1203 00:07:38.866308 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/bd03a758-0ca9-47d8-86f7-18b2d610df43-metrics-certs\") pod \"network-metrics-daemon-xshk4\" (UID: \"bd03a758-0ca9-47d8-86f7-18b2d610df43\") " pod="openshift-multus/network-metrics-daemon-xshk4" Dec 03 00:07:38 crc kubenswrapper[4953]: E1203 00:07:38.866515 4953 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 03 00:07:38 crc kubenswrapper[4953]: E1203 00:07:38.866633 4953 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/bd03a758-0ca9-47d8-86f7-18b2d610df43-metrics-certs podName:bd03a758-0ca9-47d8-86f7-18b2d610df43 nodeName:}" failed. No retries permitted until 2025-12-03 00:07:46.866604704 +0000 UTC m=+55.850885537 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/bd03a758-0ca9-47d8-86f7-18b2d610df43-metrics-certs") pod "network-metrics-daemon-xshk4" (UID: "bd03a758-0ca9-47d8-86f7-18b2d610df43") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 03 00:07:38 crc kubenswrapper[4953]: I1203 00:07:38.960217 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:38 crc kubenswrapper[4953]: I1203 00:07:38.960291 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:38 crc kubenswrapper[4953]: I1203 00:07:38.960304 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:38 crc kubenswrapper[4953]: I1203 00:07:38.960331 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:38 crc kubenswrapper[4953]: I1203 00:07:38.960348 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:38Z","lastTransitionTime":"2025-12-03T00:07:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:39 crc kubenswrapper[4953]: I1203 00:07:39.063607 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:39 crc kubenswrapper[4953]: I1203 00:07:39.064045 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:39 crc kubenswrapper[4953]: I1203 00:07:39.064126 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:39 crc kubenswrapper[4953]: I1203 00:07:39.064193 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:39 crc kubenswrapper[4953]: I1203 00:07:39.064260 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:39Z","lastTransitionTime":"2025-12-03T00:07:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:39 crc kubenswrapper[4953]: I1203 00:07:39.169020 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:39 crc kubenswrapper[4953]: I1203 00:07:39.169083 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:39 crc kubenswrapper[4953]: I1203 00:07:39.169101 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:39 crc kubenswrapper[4953]: I1203 00:07:39.169134 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:39 crc kubenswrapper[4953]: I1203 00:07:39.169153 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:39Z","lastTransitionTime":"2025-12-03T00:07:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:39 crc kubenswrapper[4953]: I1203 00:07:39.272577 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:39 crc kubenswrapper[4953]: I1203 00:07:39.272652 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:39 crc kubenswrapper[4953]: I1203 00:07:39.272672 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:39 crc kubenswrapper[4953]: I1203 00:07:39.272698 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:39 crc kubenswrapper[4953]: I1203 00:07:39.272719 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:39Z","lastTransitionTime":"2025-12-03T00:07:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:39 crc kubenswrapper[4953]: I1203 00:07:39.347745 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xshk4" Dec 03 00:07:39 crc kubenswrapper[4953]: E1203 00:07:39.348055 4953 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xshk4" podUID="bd03a758-0ca9-47d8-86f7-18b2d610df43" Dec 03 00:07:39 crc kubenswrapper[4953]: I1203 00:07:39.376393 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:39 crc kubenswrapper[4953]: I1203 00:07:39.376443 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:39 crc kubenswrapper[4953]: I1203 00:07:39.376456 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:39 crc kubenswrapper[4953]: I1203 00:07:39.376474 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:39 crc kubenswrapper[4953]: I1203 00:07:39.376487 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:39Z","lastTransitionTime":"2025-12-03T00:07:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:39 crc kubenswrapper[4953]: I1203 00:07:39.479915 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:39 crc kubenswrapper[4953]: I1203 00:07:39.480006 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:39 crc kubenswrapper[4953]: I1203 00:07:39.480025 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:39 crc kubenswrapper[4953]: I1203 00:07:39.480055 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:39 crc kubenswrapper[4953]: I1203 00:07:39.480078 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:39Z","lastTransitionTime":"2025-12-03T00:07:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:39 crc kubenswrapper[4953]: I1203 00:07:39.583517 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:39 crc kubenswrapper[4953]: I1203 00:07:39.583580 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:39 crc kubenswrapper[4953]: I1203 00:07:39.583604 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:39 crc kubenswrapper[4953]: I1203 00:07:39.583686 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:39 crc kubenswrapper[4953]: I1203 00:07:39.583713 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:39Z","lastTransitionTime":"2025-12-03T00:07:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:39 crc kubenswrapper[4953]: I1203 00:07:39.686592 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:39 crc kubenswrapper[4953]: I1203 00:07:39.686645 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:39 crc kubenswrapper[4953]: I1203 00:07:39.686667 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:39 crc kubenswrapper[4953]: I1203 00:07:39.686693 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:39 crc kubenswrapper[4953]: I1203 00:07:39.686713 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:39Z","lastTransitionTime":"2025-12-03T00:07:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:39 crc kubenswrapper[4953]: I1203 00:07:39.789214 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:39 crc kubenswrapper[4953]: I1203 00:07:39.789267 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:39 crc kubenswrapper[4953]: I1203 00:07:39.789284 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:39 crc kubenswrapper[4953]: I1203 00:07:39.789307 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:39 crc kubenswrapper[4953]: I1203 00:07:39.789324 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:39Z","lastTransitionTime":"2025-12-03T00:07:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:39 crc kubenswrapper[4953]: I1203 00:07:39.892395 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:39 crc kubenswrapper[4953]: I1203 00:07:39.892770 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:39 crc kubenswrapper[4953]: I1203 00:07:39.892901 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:39 crc kubenswrapper[4953]: I1203 00:07:39.893100 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:39 crc kubenswrapper[4953]: I1203 00:07:39.893229 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:39Z","lastTransitionTime":"2025-12-03T00:07:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:39 crc kubenswrapper[4953]: I1203 00:07:39.996529 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:39 crc kubenswrapper[4953]: I1203 00:07:39.996574 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:39 crc kubenswrapper[4953]: I1203 00:07:39.996590 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:39 crc kubenswrapper[4953]: I1203 00:07:39.996611 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:39 crc kubenswrapper[4953]: I1203 00:07:39.996628 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:39Z","lastTransitionTime":"2025-12-03T00:07:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:40 crc kubenswrapper[4953]: I1203 00:07:40.098800 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:40 crc kubenswrapper[4953]: I1203 00:07:40.098863 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:40 crc kubenswrapper[4953]: I1203 00:07:40.098880 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:40 crc kubenswrapper[4953]: I1203 00:07:40.098902 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:40 crc kubenswrapper[4953]: I1203 00:07:40.098921 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:40Z","lastTransitionTime":"2025-12-03T00:07:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:40 crc kubenswrapper[4953]: I1203 00:07:40.201632 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:40 crc kubenswrapper[4953]: I1203 00:07:40.201688 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:40 crc kubenswrapper[4953]: I1203 00:07:40.201764 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:40 crc kubenswrapper[4953]: I1203 00:07:40.201797 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:40 crc kubenswrapper[4953]: I1203 00:07:40.201818 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:40Z","lastTransitionTime":"2025-12-03T00:07:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:40 crc kubenswrapper[4953]: I1203 00:07:40.305641 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:40 crc kubenswrapper[4953]: I1203 00:07:40.305706 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:40 crc kubenswrapper[4953]: I1203 00:07:40.305728 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:40 crc kubenswrapper[4953]: I1203 00:07:40.305768 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:40 crc kubenswrapper[4953]: I1203 00:07:40.305790 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:40Z","lastTransitionTime":"2025-12-03T00:07:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:40 crc kubenswrapper[4953]: I1203 00:07:40.348141 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 00:07:40 crc kubenswrapper[4953]: I1203 00:07:40.348141 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 00:07:40 crc kubenswrapper[4953]: E1203 00:07:40.348366 4953 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 00:07:40 crc kubenswrapper[4953]: I1203 00:07:40.348141 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 00:07:40 crc kubenswrapper[4953]: E1203 00:07:40.348481 4953 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 00:07:40 crc kubenswrapper[4953]: E1203 00:07:40.348638 4953 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 00:07:40 crc kubenswrapper[4953]: I1203 00:07:40.409613 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:40 crc kubenswrapper[4953]: I1203 00:07:40.409671 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:40 crc kubenswrapper[4953]: I1203 00:07:40.409684 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:40 crc kubenswrapper[4953]: I1203 00:07:40.409707 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:40 crc kubenswrapper[4953]: I1203 00:07:40.409721 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:40Z","lastTransitionTime":"2025-12-03T00:07:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:40 crc kubenswrapper[4953]: I1203 00:07:40.513113 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:40 crc kubenswrapper[4953]: I1203 00:07:40.513165 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:40 crc kubenswrapper[4953]: I1203 00:07:40.513177 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:40 crc kubenswrapper[4953]: I1203 00:07:40.513195 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:40 crc kubenswrapper[4953]: I1203 00:07:40.513210 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:40Z","lastTransitionTime":"2025-12-03T00:07:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:40 crc kubenswrapper[4953]: I1203 00:07:40.615685 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:40 crc kubenswrapper[4953]: I1203 00:07:40.615745 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:40 crc kubenswrapper[4953]: I1203 00:07:40.615760 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:40 crc kubenswrapper[4953]: I1203 00:07:40.615781 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:40 crc kubenswrapper[4953]: I1203 00:07:40.615794 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:40Z","lastTransitionTime":"2025-12-03T00:07:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:40 crc kubenswrapper[4953]: I1203 00:07:40.719129 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:40 crc kubenswrapper[4953]: I1203 00:07:40.719208 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:40 crc kubenswrapper[4953]: I1203 00:07:40.719234 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:40 crc kubenswrapper[4953]: I1203 00:07:40.719265 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:40 crc kubenswrapper[4953]: I1203 00:07:40.719292 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:40Z","lastTransitionTime":"2025-12-03T00:07:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:40 crc kubenswrapper[4953]: I1203 00:07:40.822155 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:40 crc kubenswrapper[4953]: I1203 00:07:40.822227 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:40 crc kubenswrapper[4953]: I1203 00:07:40.822245 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:40 crc kubenswrapper[4953]: I1203 00:07:40.822270 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:40 crc kubenswrapper[4953]: I1203 00:07:40.822292 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:40Z","lastTransitionTime":"2025-12-03T00:07:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:40 crc kubenswrapper[4953]: I1203 00:07:40.925207 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:40 crc kubenswrapper[4953]: I1203 00:07:40.925279 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:40 crc kubenswrapper[4953]: I1203 00:07:40.925298 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:40 crc kubenswrapper[4953]: I1203 00:07:40.925320 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:40 crc kubenswrapper[4953]: I1203 00:07:40.925338 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:40Z","lastTransitionTime":"2025-12-03T00:07:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:41 crc kubenswrapper[4953]: I1203 00:07:41.028760 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:41 crc kubenswrapper[4953]: I1203 00:07:41.028819 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:41 crc kubenswrapper[4953]: I1203 00:07:41.028836 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:41 crc kubenswrapper[4953]: I1203 00:07:41.028859 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:41 crc kubenswrapper[4953]: I1203 00:07:41.028875 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:41Z","lastTransitionTime":"2025-12-03T00:07:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:41 crc kubenswrapper[4953]: I1203 00:07:41.132402 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:41 crc kubenswrapper[4953]: I1203 00:07:41.132476 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:41 crc kubenswrapper[4953]: I1203 00:07:41.132493 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:41 crc kubenswrapper[4953]: I1203 00:07:41.132517 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:41 crc kubenswrapper[4953]: I1203 00:07:41.132535 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:41Z","lastTransitionTime":"2025-12-03T00:07:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:41 crc kubenswrapper[4953]: I1203 00:07:41.235478 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:41 crc kubenswrapper[4953]: I1203 00:07:41.235574 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:41 crc kubenswrapper[4953]: I1203 00:07:41.235591 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:41 crc kubenswrapper[4953]: I1203 00:07:41.235620 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:41 crc kubenswrapper[4953]: I1203 00:07:41.235643 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:41Z","lastTransitionTime":"2025-12-03T00:07:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:41 crc kubenswrapper[4953]: I1203 00:07:41.339349 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:41 crc kubenswrapper[4953]: I1203 00:07:41.339432 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:41 crc kubenswrapper[4953]: I1203 00:07:41.339457 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:41 crc kubenswrapper[4953]: I1203 00:07:41.339491 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:41 crc kubenswrapper[4953]: I1203 00:07:41.339515 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:41Z","lastTransitionTime":"2025-12-03T00:07:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:41 crc kubenswrapper[4953]: I1203 00:07:41.349350 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xshk4" Dec 03 00:07:41 crc kubenswrapper[4953]: E1203 00:07:41.349559 4953 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xshk4" podUID="bd03a758-0ca9-47d8-86f7-18b2d610df43" Dec 03 00:07:41 crc kubenswrapper[4953]: I1203 00:07:41.373480 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:41Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:41 crc kubenswrapper[4953]: I1203 00:07:41.393869 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2ed9c7fa96c8951fbc1728fe7a91fefdd999077285621df9c1abedf9ae0c449\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:41Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:41 crc kubenswrapper[4953]: I1203 00:07:41.415049 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-675qk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"43a4bad8-71f4-441a-8302-aeb526a986af\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5b5794fef6477aab5deed8f642e8cc50b37eb1fd0cc6f19c003f34ad31c1e3c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gdrw4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:07:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-675qk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:41Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:41 crc kubenswrapper[4953]: I1203 00:07:41.436376 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:10Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:41Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:41 crc kubenswrapper[4953]: I1203 00:07:41.442596 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:41 crc kubenswrapper[4953]: I1203 00:07:41.442717 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:41 crc kubenswrapper[4953]: I1203 00:07:41.442741 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:41 crc kubenswrapper[4953]: I1203 00:07:41.442812 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:41 crc kubenswrapper[4953]: I1203 00:07:41.442838 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:41Z","lastTransitionTime":"2025-12-03T00:07:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:41 crc kubenswrapper[4953]: I1203 00:07:41.460781 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09657a9f448a4c92a759c73a0a892076d7eb46393e6fd565e5f73f34c71aac23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5af5eac4a58870f63497f99b503cfa5fb922278095e0332ec50376925ecf2109\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:41Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:41 crc kubenswrapper[4953]: I1203 00:07:41.487589 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-wj6rz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"96aa67a0-15dd-420d-a1d5-ed239d7df1fb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01815cfd3f58c3f3f5f2a894c1790e0b69c5060c48dc0b1deff050b6ad7af916\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kf62m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b12bfb5bdf9fd4f2c50a335fc82df6ee0440eaf8741897da03aa2dae368cc6b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b12bfb5bdf9fd4f2c50a335fc82df6ee0440eaf8741897da03aa2dae368cc6b4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:07:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kf62m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83aaa5fa28110bc0ef05aff75cd1442bc9bbdd33b822ae2ef22d4e4ab83238bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://83aaa5fa28110bc0ef05aff75cd1442bc9bbdd33b822ae2ef22d4e4ab83238bc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:07:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:07:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kf62m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ab6213a33a19fedebce6d74a0ebbb3b3a2de5687504e21cc24093cc42ade635\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ab6213a33a19fedebce6d74a0ebbb3b3a2de5687504e21cc24093cc42ade635\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:07:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:07:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kf62m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://de480488843afa6b8dede1e1b42b5181dca1993e2f39dc1b461a18a57e3668e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://de480488843afa6b8dede1e1b42b5181dca1993e2f39dc1b461a18a57e3668e1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:07:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kf62m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://97be4ba692bc61078d4b0ba39fb0c12772548c6c9251ff3356ea702f151223cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://97be4ba692bc61078d4b0ba39fb0c12772548c6c9251ff3356ea702f151223cd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:07:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kf62m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1a1f009fcbd8af9217db1dc97f662719f8c1db232df0e2804e44c881abb1df2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b1a1f009fcbd8af9217db1dc97f662719f8c1db232df0e2804e44c881abb1df2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:07:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:07:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kf62m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:07:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-wj6rz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:41Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:41 crc kubenswrapper[4953]: I1203 00:07:41.500078 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-kl5s9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d85e5c97-16ad-4bd7-9e73-b7f12678be97\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c21df91a9e4e3165686882b2eb559498224063dc633c3a63f958598108eca29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j24lb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:07:16Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-kl5s9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:41Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:41 crc kubenswrapper[4953]: I1203 00:07:41.513667 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f94f8c51-c665-4ca3-bf17-7a04ea5c7ade\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72c5cd7dc1c3a8adfedab6ba43a59d80b3efe65f2bb56be95e76cc86a409dcf3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2810ffc1b141fa7da811406053db9d72c708edf5f2bb1dd75e4d52ab35fd24c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c67b4277b20a9f1287f50ef82d0785c9d02543a8e7ed26e0413eae332b866db7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://258d78e1ad9d3dcda6976909ae3fe71aecead5b995fc66b3fd6e2722d80b5ab8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:06:51Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:41Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:41 crc kubenswrapper[4953]: I1203 00:07:41.530839 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:10Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:41Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:41 crc kubenswrapper[4953]: I1203 00:07:41.546444 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:41 crc kubenswrapper[4953]: I1203 00:07:41.546501 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:41 crc kubenswrapper[4953]: I1203 00:07:41.546523 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:41 crc kubenswrapper[4953]: I1203 00:07:41.546547 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:41 crc kubenswrapper[4953]: I1203 00:07:41.546565 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:41Z","lastTransitionTime":"2025-12-03T00:07:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:41 crc kubenswrapper[4953]: I1203 00:07:41.552678 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce8b414e1fc22eb30decb6a6411b672e7f443055cb89ebb029fbf6ffc1d5dca7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:41Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:41 crc kubenswrapper[4953]: I1203 00:07:41.568243 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-6dfbn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53ffb1e9-9a2e-4869-80d7-776f4aa7ef8b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ab4205c0d38ad6a0c32077eadb1fd5fd3a279d9c8b8a2ffbe693ec2c4018370c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ct26c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7df84405eda520ad5a22d35750edf07ef3c3977a756ad5739c54f91cba9da7b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ct26c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:07:16Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-6dfbn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:41Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:41 crc kubenswrapper[4953]: I1203 00:07:41.589462 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-txffx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8816ebee-c735-4b8d-b1ec-65349707b347\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:16Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:16Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4b85224419a4075acb340e4b11095f2adb5c316e285afa4e19295c2181c0d24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0f6f645a58993f052a6f365bb808f62d99f7fea74886c5cc022b2d86112321a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a61b951a9cbab2c48c0f2df8351f97d7ff71bf3674f641fe5eac30155232b224\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd5f428de229749485267e81d34e31b708728b52474032e88460e54747f21e11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://00a2ea4a9bb870fce30e7fa845f787f184e7a19e42c95511655ea8c5f84a26d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc0a5d4cc9d614d8492626a1b114bdc52f510eb5965d842c02c083f08d4849a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://efbfc74a8598c47a658c06b597fa36d2a5ac8d02265d2864fd2832dd11ecb2cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aecc0ae6dd34893c9023da3b94ca01ae96ac1e7cd7f5fd18748cce6c273a7f9e\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T00:07:26Z\\\",\\\"message\\\":\\\"ller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1203 00:07:25.992760 6257 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1203 00:07:25.992789 6257 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1203 00:07:25.992837 6257 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1203 00:07:25.992889 6257 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1203 00:07:25.992927 6257 handler.go:208] Removed *v1.Node event handler 7\\\\nI1203 00:07:25.992967 6257 handler.go:208] Removed *v1.Node event handler 2\\\\nI1203 00:07:25.992993 6257 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1203 00:07:25.992947 6257 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1203 00:07:25.993079 6257 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1203 00:07:25.992969 6257 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1203 00:07:25.993139 6257 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1203 00:07:25.993168 6257 factory.go:656] Stopping watch factory\\\\nI1203 00:07:25.993190 6257 ovnkube.go:599] Stopped ovnkube\\\\nI1203 00:07:25.993215 6257 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1203 00:07:25.993246 6257 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T00:07:23Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://efbfc74a8598c47a658c06b597fa36d2a5ac8d02265d2864fd2832dd11ecb2cc\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T00:07:29Z\\\",\\\"message\\\":\\\" 6400 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1203 00:07:29.614806 6400 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1203 00:07:29.614821 6400 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1203 00:07:29.614822 6400 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 00:07:29.614854 6400 factory.go:656] Stopping watch factory\\\\nI1203 00:07:29.614869 6400 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1203 00:07:29.614878 6400 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1203 00:07:29.614875 6400 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1203 00:07:29.614894 6400 handler.go:208] Removed *v1.Node event handler 7\\\\nI1203 00:07:29.614906 6400 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1203 00:07:29.614915 6400 reflector.go:311] Stopping reflector *v1.NetworkAttachmentDefinition (0s) from github.com/k8snetworkplumbingwg/network-attachment-definition-client/pkg/client/informers/externalversions/factory.go:117\\\\nI1203 00:07:29.615144 6400 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 00:07:29.615318 6400 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T00:07:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7deadc8e135ff1bc7c5a2f093e489bb5b4b6f166bc874c6acc2a841c5397f11e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af4c7764fca87144d01f3738ea010308744ccf50c54b7c471ccc213112de00d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af4c7764fca87144d01f3738ea010308744ccf50c54b7c471ccc213112de00d5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:07:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:07:16Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-txffx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:41Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:41 crc kubenswrapper[4953]: I1203 00:07:41.604113 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5cmlc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7e11a1d3-ecee-4bf0-87bf-085fab97b73d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d42390b9c53efc63add8ddcebf03c976f5e5448eaa2c34e076b2dc90dbac60fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6x2sc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:07:18Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5cmlc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:41Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:41 crc kubenswrapper[4953]: I1203 00:07:41.618389 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-chmnz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a12fe6d-d25a-4b08-90da-cc9e510980a7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://711ab839cb9c0a90554dadbd45d810fc78aac09739340482fef8de8743b77283\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c982g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f802d5628a8cc32618044b9383211d5a492c56060d427f953b7793a1064029e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c982g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:07:29Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-chmnz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:41Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:41 crc kubenswrapper[4953]: I1203 00:07:41.629165 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-xshk4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bd03a758-0ca9-47d8-86f7-18b2d610df43\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47gfh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47gfh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:07:31Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-xshk4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:41Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:41 crc kubenswrapper[4953]: I1203 00:07:41.644034 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cf821840-5353-4620-8f73-99b56b371ab0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cae5e2994fcc24dc9b2ea73e79572615b3311edc42642b7f278fb84123989b23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://29e4b9d1b70a93fb133e0759ad148ae133766a66efeae3ff06a21dfae0e805ff\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dec43536772aa7c3f5a662fa7f3bd73b1cae2791ad454f85792287828d0ffeed\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://034e0bac6892d361700b8f33d6427b8a55e2bd3f89b14ea77a374c6be94d25e2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://761973393a93c59c23e69f41b8ac0d16abc35abd61ecbc6d51a42d94cd664233\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T00:07:09Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 00:07:03.713374 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 00:07:03.715052 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1528563048/tls.crt::/tmp/serving-cert-1528563048/tls.key\\\\\\\"\\\\nI1203 00:07:09.300052 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 00:07:09.304192 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 00:07:09.304255 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 00:07:09.304309 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 00:07:09.304323 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 00:07:09.313563 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 00:07:09.313597 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 00:07:09.313603 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 00:07:09.313609 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 00:07:09.313613 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 00:07:09.313618 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 00:07:09.313622 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 00:07:09.313709 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 00:07:09.327347 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T00:06:53Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c143d4232fb8bf7ef5999a664e431a00ab0db030f4d31a28d0bb21e00e38581\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:53Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72af6c27296eae2e275d455795c68f0d272869d63642b8353b17d466d0c901cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://72af6c27296eae2e275d455795c68f0d272869d63642b8353b17d466d0c901cf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:06:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:06:51Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:41Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:41 crc kubenswrapper[4953]: I1203 00:07:41.649786 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:41 crc kubenswrapper[4953]: I1203 00:07:41.649838 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:41 crc kubenswrapper[4953]: I1203 00:07:41.649857 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:41 crc kubenswrapper[4953]: I1203 00:07:41.649878 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:41 crc kubenswrapper[4953]: I1203 00:07:41.649894 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:41Z","lastTransitionTime":"2025-12-03T00:07:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:41 crc kubenswrapper[4953]: I1203 00:07:41.753071 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:41 crc kubenswrapper[4953]: I1203 00:07:41.753129 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:41 crc kubenswrapper[4953]: I1203 00:07:41.753146 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:41 crc kubenswrapper[4953]: I1203 00:07:41.753170 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:41 crc kubenswrapper[4953]: I1203 00:07:41.753188 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:41Z","lastTransitionTime":"2025-12-03T00:07:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:41 crc kubenswrapper[4953]: I1203 00:07:41.855482 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:41 crc kubenswrapper[4953]: I1203 00:07:41.855554 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:41 crc kubenswrapper[4953]: I1203 00:07:41.855578 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:41 crc kubenswrapper[4953]: I1203 00:07:41.855604 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:41 crc kubenswrapper[4953]: I1203 00:07:41.855628 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:41Z","lastTransitionTime":"2025-12-03T00:07:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:41 crc kubenswrapper[4953]: I1203 00:07:41.958410 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:41 crc kubenswrapper[4953]: I1203 00:07:41.958503 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:41 crc kubenswrapper[4953]: I1203 00:07:41.958530 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:41 crc kubenswrapper[4953]: I1203 00:07:41.958562 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:41 crc kubenswrapper[4953]: I1203 00:07:41.958588 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:41Z","lastTransitionTime":"2025-12-03T00:07:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:42 crc kubenswrapper[4953]: I1203 00:07:42.061452 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:42 crc kubenswrapper[4953]: I1203 00:07:42.061534 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:42 crc kubenswrapper[4953]: I1203 00:07:42.061556 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:42 crc kubenswrapper[4953]: I1203 00:07:42.061580 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:42 crc kubenswrapper[4953]: I1203 00:07:42.061602 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:42Z","lastTransitionTime":"2025-12-03T00:07:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:42 crc kubenswrapper[4953]: I1203 00:07:42.164722 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:42 crc kubenswrapper[4953]: I1203 00:07:42.164807 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:42 crc kubenswrapper[4953]: I1203 00:07:42.164829 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:42 crc kubenswrapper[4953]: I1203 00:07:42.164857 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:42 crc kubenswrapper[4953]: I1203 00:07:42.164881 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:42Z","lastTransitionTime":"2025-12-03T00:07:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:42 crc kubenswrapper[4953]: I1203 00:07:42.204027 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 00:07:42 crc kubenswrapper[4953]: I1203 00:07:42.204206 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 00:07:42 crc kubenswrapper[4953]: E1203 00:07:42.204275 4953 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 00:08:14.204238867 +0000 UTC m=+83.188519690 (durationBeforeRetry 32s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 00:07:42 crc kubenswrapper[4953]: I1203 00:07:42.204369 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 00:07:42 crc kubenswrapper[4953]: E1203 00:07:42.204407 4953 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 03 00:07:42 crc kubenswrapper[4953]: E1203 00:07:42.204433 4953 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 03 00:07:42 crc kubenswrapper[4953]: E1203 00:07:42.204456 4953 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 00:07:42 crc kubenswrapper[4953]: I1203 00:07:42.204460 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 00:07:42 crc kubenswrapper[4953]: I1203 00:07:42.204512 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 00:07:42 crc kubenswrapper[4953]: E1203 00:07:42.204613 4953 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-03 00:08:14.204597576 +0000 UTC m=+83.188878399 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 00:07:42 crc kubenswrapper[4953]: E1203 00:07:42.204687 4953 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 03 00:07:42 crc kubenswrapper[4953]: E1203 00:07:42.204719 4953 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 03 00:07:42 crc kubenswrapper[4953]: E1203 00:07:42.204737 4953 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 03 00:07:42 crc kubenswrapper[4953]: E1203 00:07:42.204764 4953 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 03 00:07:42 crc kubenswrapper[4953]: E1203 00:07:42.204808 4953 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 00:07:42 crc kubenswrapper[4953]: E1203 00:07:42.204753 4953 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-03 00:08:14.204739649 +0000 UTC m=+83.189020472 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 03 00:07:42 crc kubenswrapper[4953]: E1203 00:07:42.204916 4953 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-03 00:08:14.204886253 +0000 UTC m=+83.189167066 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 03 00:07:42 crc kubenswrapper[4953]: E1203 00:07:42.204959 4953 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-03 00:08:14.204947324 +0000 UTC m=+83.189228147 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 00:07:42 crc kubenswrapper[4953]: I1203 00:07:42.268268 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:42 crc kubenswrapper[4953]: I1203 00:07:42.268323 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:42 crc kubenswrapper[4953]: I1203 00:07:42.268382 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:42 crc kubenswrapper[4953]: I1203 00:07:42.268406 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:42 crc kubenswrapper[4953]: I1203 00:07:42.268424 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:42Z","lastTransitionTime":"2025-12-03T00:07:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:42 crc kubenswrapper[4953]: I1203 00:07:42.347912 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 00:07:42 crc kubenswrapper[4953]: I1203 00:07:42.347912 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 00:07:42 crc kubenswrapper[4953]: I1203 00:07:42.348060 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 00:07:42 crc kubenswrapper[4953]: E1203 00:07:42.348136 4953 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 00:07:42 crc kubenswrapper[4953]: E1203 00:07:42.348276 4953 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 00:07:42 crc kubenswrapper[4953]: E1203 00:07:42.348481 4953 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 00:07:42 crc kubenswrapper[4953]: I1203 00:07:42.372154 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:42 crc kubenswrapper[4953]: I1203 00:07:42.372205 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:42 crc kubenswrapper[4953]: I1203 00:07:42.372223 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:42 crc kubenswrapper[4953]: I1203 00:07:42.372246 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:42 crc kubenswrapper[4953]: I1203 00:07:42.372265 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:42Z","lastTransitionTime":"2025-12-03T00:07:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:42 crc kubenswrapper[4953]: I1203 00:07:42.378842 4953 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 03 00:07:42 crc kubenswrapper[4953]: I1203 00:07:42.395760 4953 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler/openshift-kube-scheduler-crc"] Dec 03 00:07:42 crc kubenswrapper[4953]: I1203 00:07:42.407374 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f94f8c51-c665-4ca3-bf17-7a04ea5c7ade\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72c5cd7dc1c3a8adfedab6ba43a59d80b3efe65f2bb56be95e76cc86a409dcf3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2810ffc1b141fa7da811406053db9d72c708edf5f2bb1dd75e4d52ab35fd24c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c67b4277b20a9f1287f50ef82d0785c9d02543a8e7ed26e0413eae332b866db7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://258d78e1ad9d3dcda6976909ae3fe71aecead5b995fc66b3fd6e2722d80b5ab8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:06:51Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:42Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:42 crc kubenswrapper[4953]: I1203 00:07:42.427726 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09657a9f448a4c92a759c73a0a892076d7eb46393e6fd565e5f73f34c71aac23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5af5eac4a58870f63497f99b503cfa5fb922278095e0332ec50376925ecf2109\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:42Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:42 crc kubenswrapper[4953]: I1203 00:07:42.450962 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-wj6rz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"96aa67a0-15dd-420d-a1d5-ed239d7df1fb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01815cfd3f58c3f3f5f2a894c1790e0b69c5060c48dc0b1deff050b6ad7af916\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kf62m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b12bfb5bdf9fd4f2c50a335fc82df6ee0440eaf8741897da03aa2dae368cc6b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b12bfb5bdf9fd4f2c50a335fc82df6ee0440eaf8741897da03aa2dae368cc6b4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:07:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kf62m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83aaa5fa28110bc0ef05aff75cd1442bc9bbdd33b822ae2ef22d4e4ab83238bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://83aaa5fa28110bc0ef05aff75cd1442bc9bbdd33b822ae2ef22d4e4ab83238bc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:07:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:07:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kf62m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ab6213a33a19fedebce6d74a0ebbb3b3a2de5687504e21cc24093cc42ade635\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ab6213a33a19fedebce6d74a0ebbb3b3a2de5687504e21cc24093cc42ade635\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:07:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:07:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kf62m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://de480488843afa6b8dede1e1b42b5181dca1993e2f39dc1b461a18a57e3668e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://de480488843afa6b8dede1e1b42b5181dca1993e2f39dc1b461a18a57e3668e1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:07:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kf62m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://97be4ba692bc61078d4b0ba39fb0c12772548c6c9251ff3356ea702f151223cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://97be4ba692bc61078d4b0ba39fb0c12772548c6c9251ff3356ea702f151223cd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:07:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kf62m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1a1f009fcbd8af9217db1dc97f662719f8c1db232df0e2804e44c881abb1df2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b1a1f009fcbd8af9217db1dc97f662719f8c1db232df0e2804e44c881abb1df2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:07:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:07:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kf62m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:07:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-wj6rz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:42Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:42 crc kubenswrapper[4953]: I1203 00:07:42.466932 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-kl5s9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d85e5c97-16ad-4bd7-9e73-b7f12678be97\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c21df91a9e4e3165686882b2eb559498224063dc633c3a63f958598108eca29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j24lb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:07:16Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-kl5s9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:42Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:42 crc kubenswrapper[4953]: I1203 00:07:42.475337 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:42 crc kubenswrapper[4953]: I1203 00:07:42.475398 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:42 crc kubenswrapper[4953]: I1203 00:07:42.475416 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:42 crc kubenswrapper[4953]: I1203 00:07:42.475443 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:42 crc kubenswrapper[4953]: I1203 00:07:42.475464 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:42Z","lastTransitionTime":"2025-12-03T00:07:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:42 crc kubenswrapper[4953]: I1203 00:07:42.495432 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cf821840-5353-4620-8f73-99b56b371ab0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cae5e2994fcc24dc9b2ea73e79572615b3311edc42642b7f278fb84123989b23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://29e4b9d1b70a93fb133e0759ad148ae133766a66efeae3ff06a21dfae0e805ff\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dec43536772aa7c3f5a662fa7f3bd73b1cae2791ad454f85792287828d0ffeed\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://034e0bac6892d361700b8f33d6427b8a55e2bd3f89b14ea77a374c6be94d25e2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://761973393a93c59c23e69f41b8ac0d16abc35abd61ecbc6d51a42d94cd664233\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T00:07:09Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 00:07:03.713374 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 00:07:03.715052 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1528563048/tls.crt::/tmp/serving-cert-1528563048/tls.key\\\\\\\"\\\\nI1203 00:07:09.300052 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 00:07:09.304192 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 00:07:09.304255 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 00:07:09.304309 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 00:07:09.304323 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 00:07:09.313563 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 00:07:09.313597 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 00:07:09.313603 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 00:07:09.313609 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 00:07:09.313613 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 00:07:09.313618 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 00:07:09.313622 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 00:07:09.313709 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 00:07:09.327347 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T00:06:53Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c143d4232fb8bf7ef5999a664e431a00ab0db030f4d31a28d0bb21e00e38581\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:53Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72af6c27296eae2e275d455795c68f0d272869d63642b8353b17d466d0c901cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://72af6c27296eae2e275d455795c68f0d272869d63642b8353b17d466d0c901cf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:06:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:06:51Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:42Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:42 crc kubenswrapper[4953]: I1203 00:07:42.516287 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:10Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:42Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:42 crc kubenswrapper[4953]: I1203 00:07:42.536614 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce8b414e1fc22eb30decb6a6411b672e7f443055cb89ebb029fbf6ffc1d5dca7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:42Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:42 crc kubenswrapper[4953]: I1203 00:07:42.553433 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-6dfbn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53ffb1e9-9a2e-4869-80d7-776f4aa7ef8b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ab4205c0d38ad6a0c32077eadb1fd5fd3a279d9c8b8a2ffbe693ec2c4018370c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ct26c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7df84405eda520ad5a22d35750edf07ef3c3977a756ad5739c54f91cba9da7b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ct26c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:07:16Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-6dfbn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:42Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:42 crc kubenswrapper[4953]: I1203 00:07:42.579253 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:42 crc kubenswrapper[4953]: I1203 00:07:42.579313 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:42 crc kubenswrapper[4953]: I1203 00:07:42.579331 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:42 crc kubenswrapper[4953]: I1203 00:07:42.579355 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:42 crc kubenswrapper[4953]: I1203 00:07:42.579372 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:42Z","lastTransitionTime":"2025-12-03T00:07:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:42 crc kubenswrapper[4953]: I1203 00:07:42.586218 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-txffx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8816ebee-c735-4b8d-b1ec-65349707b347\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:16Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:16Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4b85224419a4075acb340e4b11095f2adb5c316e285afa4e19295c2181c0d24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0f6f645a58993f052a6f365bb808f62d99f7fea74886c5cc022b2d86112321a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a61b951a9cbab2c48c0f2df8351f97d7ff71bf3674f641fe5eac30155232b224\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd5f428de229749485267e81d34e31b708728b52474032e88460e54747f21e11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://00a2ea4a9bb870fce30e7fa845f787f184e7a19e42c95511655ea8c5f84a26d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc0a5d4cc9d614d8492626a1b114bdc52f510eb5965d842c02c083f08d4849a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://efbfc74a8598c47a658c06b597fa36d2a5ac8d02265d2864fd2832dd11ecb2cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aecc0ae6dd34893c9023da3b94ca01ae96ac1e7cd7f5fd18748cce6c273a7f9e\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T00:07:26Z\\\",\\\"message\\\":\\\"ller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1203 00:07:25.992760 6257 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1203 00:07:25.992789 6257 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1203 00:07:25.992837 6257 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1203 00:07:25.992889 6257 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1203 00:07:25.992927 6257 handler.go:208] Removed *v1.Node event handler 7\\\\nI1203 00:07:25.992967 6257 handler.go:208] Removed *v1.Node event handler 2\\\\nI1203 00:07:25.992993 6257 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1203 00:07:25.992947 6257 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1203 00:07:25.993079 6257 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1203 00:07:25.992969 6257 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1203 00:07:25.993139 6257 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1203 00:07:25.993168 6257 factory.go:656] Stopping watch factory\\\\nI1203 00:07:25.993190 6257 ovnkube.go:599] Stopped ovnkube\\\\nI1203 00:07:25.993215 6257 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1203 00:07:25.993246 6257 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T00:07:23Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://efbfc74a8598c47a658c06b597fa36d2a5ac8d02265d2864fd2832dd11ecb2cc\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T00:07:29Z\\\",\\\"message\\\":\\\" 6400 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1203 00:07:29.614806 6400 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1203 00:07:29.614821 6400 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1203 00:07:29.614822 6400 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 00:07:29.614854 6400 factory.go:656] Stopping watch factory\\\\nI1203 00:07:29.614869 6400 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1203 00:07:29.614878 6400 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1203 00:07:29.614875 6400 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1203 00:07:29.614894 6400 handler.go:208] Removed *v1.Node event handler 7\\\\nI1203 00:07:29.614906 6400 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1203 00:07:29.614915 6400 reflector.go:311] Stopping reflector *v1.NetworkAttachmentDefinition (0s) from github.com/k8snetworkplumbingwg/network-attachment-definition-client/pkg/client/informers/externalversions/factory.go:117\\\\nI1203 00:07:29.615144 6400 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 00:07:29.615318 6400 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T00:07:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7deadc8e135ff1bc7c5a2f093e489bb5b4b6f166bc874c6acc2a841c5397f11e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af4c7764fca87144d01f3738ea010308744ccf50c54b7c471ccc213112de00d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af4c7764fca87144d01f3738ea010308744ccf50c54b7c471ccc213112de00d5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:07:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:07:16Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-txffx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:42Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:42 crc kubenswrapper[4953]: I1203 00:07:42.601911 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5cmlc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7e11a1d3-ecee-4bf0-87bf-085fab97b73d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d42390b9c53efc63add8ddcebf03c976f5e5448eaa2c34e076b2dc90dbac60fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6x2sc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:07:18Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5cmlc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:42Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:42 crc kubenswrapper[4953]: I1203 00:07:42.618103 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-chmnz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a12fe6d-d25a-4b08-90da-cc9e510980a7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://711ab839cb9c0a90554dadbd45d810fc78aac09739340482fef8de8743b77283\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c982g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f802d5628a8cc32618044b9383211d5a492c56060d427f953b7793a1064029e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c982g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:07:29Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-chmnz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:42Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:42 crc kubenswrapper[4953]: I1203 00:07:42.631882 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-xshk4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bd03a758-0ca9-47d8-86f7-18b2d610df43\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47gfh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47gfh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:07:31Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-xshk4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:42Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:42 crc kubenswrapper[4953]: I1203 00:07:42.647281 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:42Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:42 crc kubenswrapper[4953]: I1203 00:07:42.664144 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2ed9c7fa96c8951fbc1728fe7a91fefdd999077285621df9c1abedf9ae0c449\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:42Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:42 crc kubenswrapper[4953]: I1203 00:07:42.683060 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:10Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:42Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:42 crc kubenswrapper[4953]: I1203 00:07:42.683203 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:42 crc kubenswrapper[4953]: I1203 00:07:42.683242 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:42 crc kubenswrapper[4953]: I1203 00:07:42.683259 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:42 crc kubenswrapper[4953]: I1203 00:07:42.683282 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:42 crc kubenswrapper[4953]: I1203 00:07:42.683300 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:42Z","lastTransitionTime":"2025-12-03T00:07:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:42 crc kubenswrapper[4953]: I1203 00:07:42.704771 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-675qk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"43a4bad8-71f4-441a-8302-aeb526a986af\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5b5794fef6477aab5deed8f642e8cc50b37eb1fd0cc6f19c003f34ad31c1e3c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gdrw4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:07:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-675qk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:42Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:42 crc kubenswrapper[4953]: I1203 00:07:42.786466 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:42 crc kubenswrapper[4953]: I1203 00:07:42.786534 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:42 crc kubenswrapper[4953]: I1203 00:07:42.786551 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:42 crc kubenswrapper[4953]: I1203 00:07:42.786576 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:42 crc kubenswrapper[4953]: I1203 00:07:42.786594 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:42Z","lastTransitionTime":"2025-12-03T00:07:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:42 crc kubenswrapper[4953]: I1203 00:07:42.890414 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:42 crc kubenswrapper[4953]: I1203 00:07:42.890504 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:42 crc kubenswrapper[4953]: I1203 00:07:42.890530 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:42 crc kubenswrapper[4953]: I1203 00:07:42.890559 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:42 crc kubenswrapper[4953]: I1203 00:07:42.890581 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:42Z","lastTransitionTime":"2025-12-03T00:07:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:42 crc kubenswrapper[4953]: I1203 00:07:42.994079 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:42 crc kubenswrapper[4953]: I1203 00:07:42.994140 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:42 crc kubenswrapper[4953]: I1203 00:07:42.994148 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:42 crc kubenswrapper[4953]: I1203 00:07:42.994167 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:42 crc kubenswrapper[4953]: I1203 00:07:42.994182 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:42Z","lastTransitionTime":"2025-12-03T00:07:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:43 crc kubenswrapper[4953]: I1203 00:07:43.096921 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:43 crc kubenswrapper[4953]: I1203 00:07:43.097044 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:43 crc kubenswrapper[4953]: I1203 00:07:43.097063 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:43 crc kubenswrapper[4953]: I1203 00:07:43.097086 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:43 crc kubenswrapper[4953]: I1203 00:07:43.097104 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:43Z","lastTransitionTime":"2025-12-03T00:07:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:43 crc kubenswrapper[4953]: I1203 00:07:43.122721 4953 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-txffx" Dec 03 00:07:43 crc kubenswrapper[4953]: I1203 00:07:43.123472 4953 scope.go:117] "RemoveContainer" containerID="efbfc74a8598c47a658c06b597fa36d2a5ac8d02265d2864fd2832dd11ecb2cc" Dec 03 00:07:43 crc kubenswrapper[4953]: I1203 00:07:43.148961 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:43Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:43 crc kubenswrapper[4953]: I1203 00:07:43.178954 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2ed9c7fa96c8951fbc1728fe7a91fefdd999077285621df9c1abedf9ae0c449\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:43Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:43 crc kubenswrapper[4953]: I1203 00:07:43.196770 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:10Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:43Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:43 crc kubenswrapper[4953]: I1203 00:07:43.199603 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:43 crc kubenswrapper[4953]: I1203 00:07:43.199643 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:43 crc kubenswrapper[4953]: I1203 00:07:43.199656 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:43 crc kubenswrapper[4953]: I1203 00:07:43.199676 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:43 crc kubenswrapper[4953]: I1203 00:07:43.199687 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:43Z","lastTransitionTime":"2025-12-03T00:07:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:43 crc kubenswrapper[4953]: I1203 00:07:43.217127 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-675qk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"43a4bad8-71f4-441a-8302-aeb526a986af\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5b5794fef6477aab5deed8f642e8cc50b37eb1fd0cc6f19c003f34ad31c1e3c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gdrw4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:07:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-675qk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:43Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:43 crc kubenswrapper[4953]: I1203 00:07:43.229365 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f94f8c51-c665-4ca3-bf17-7a04ea5c7ade\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72c5cd7dc1c3a8adfedab6ba43a59d80b3efe65f2bb56be95e76cc86a409dcf3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2810ffc1b141fa7da811406053db9d72c708edf5f2bb1dd75e4d52ab35fd24c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c67b4277b20a9f1287f50ef82d0785c9d02543a8e7ed26e0413eae332b866db7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://258d78e1ad9d3dcda6976909ae3fe71aecead5b995fc66b3fd6e2722d80b5ab8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:06:51Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:43Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:43 crc kubenswrapper[4953]: I1203 00:07:43.244543 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5652d0f3-3fd7-4af6-8941-9119e1177641\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2030aef5b7170991268fa5dc7e7ed8ce1fbe3a2765165a10c25251426575330f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ab6052f1abf17225155ce1bd86afa742e6e78f507b7deedd7161cd8fba9120b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://650b6c79cda2e1768ba1d8763db799e210ce2748893512068d25e60f821937ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d1d60162f7f6885d4f51b4ab44163cfae6d2d04355b1dd710e26b022450005df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d1d60162f7f6885d4f51b4ab44163cfae6d2d04355b1dd710e26b022450005df\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:06:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:06:51Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:43Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:43 crc kubenswrapper[4953]: I1203 00:07:43.259336 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09657a9f448a4c92a759c73a0a892076d7eb46393e6fd565e5f73f34c71aac23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5af5eac4a58870f63497f99b503cfa5fb922278095e0332ec50376925ecf2109\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:43Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:43 crc kubenswrapper[4953]: I1203 00:07:43.284726 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-wj6rz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"96aa67a0-15dd-420d-a1d5-ed239d7df1fb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01815cfd3f58c3f3f5f2a894c1790e0b69c5060c48dc0b1deff050b6ad7af916\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kf62m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b12bfb5bdf9fd4f2c50a335fc82df6ee0440eaf8741897da03aa2dae368cc6b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b12bfb5bdf9fd4f2c50a335fc82df6ee0440eaf8741897da03aa2dae368cc6b4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:07:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kf62m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83aaa5fa28110bc0ef05aff75cd1442bc9bbdd33b822ae2ef22d4e4ab83238bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://83aaa5fa28110bc0ef05aff75cd1442bc9bbdd33b822ae2ef22d4e4ab83238bc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:07:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:07:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kf62m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ab6213a33a19fedebce6d74a0ebbb3b3a2de5687504e21cc24093cc42ade635\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ab6213a33a19fedebce6d74a0ebbb3b3a2de5687504e21cc24093cc42ade635\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:07:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:07:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kf62m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://de480488843afa6b8dede1e1b42b5181dca1993e2f39dc1b461a18a57e3668e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://de480488843afa6b8dede1e1b42b5181dca1993e2f39dc1b461a18a57e3668e1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:07:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kf62m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://97be4ba692bc61078d4b0ba39fb0c12772548c6c9251ff3356ea702f151223cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://97be4ba692bc61078d4b0ba39fb0c12772548c6c9251ff3356ea702f151223cd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:07:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kf62m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1a1f009fcbd8af9217db1dc97f662719f8c1db232df0e2804e44c881abb1df2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b1a1f009fcbd8af9217db1dc97f662719f8c1db232df0e2804e44c881abb1df2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:07:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:07:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kf62m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:07:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-wj6rz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:43Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:43 crc kubenswrapper[4953]: I1203 00:07:43.299702 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-kl5s9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d85e5c97-16ad-4bd7-9e73-b7f12678be97\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c21df91a9e4e3165686882b2eb559498224063dc633c3a63f958598108eca29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j24lb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:07:16Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-kl5s9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:43Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:43 crc kubenswrapper[4953]: I1203 00:07:43.302729 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:43 crc kubenswrapper[4953]: I1203 00:07:43.302774 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:43 crc kubenswrapper[4953]: I1203 00:07:43.302785 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:43 crc kubenswrapper[4953]: I1203 00:07:43.302806 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:43 crc kubenswrapper[4953]: I1203 00:07:43.302830 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:43Z","lastTransitionTime":"2025-12-03T00:07:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:43 crc kubenswrapper[4953]: I1203 00:07:43.314475 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5cmlc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7e11a1d3-ecee-4bf0-87bf-085fab97b73d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d42390b9c53efc63add8ddcebf03c976f5e5448eaa2c34e076b2dc90dbac60fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6x2sc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:07:18Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5cmlc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:43Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:43 crc kubenswrapper[4953]: I1203 00:07:43.330662 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-chmnz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a12fe6d-d25a-4b08-90da-cc9e510980a7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://711ab839cb9c0a90554dadbd45d810fc78aac09739340482fef8de8743b77283\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c982g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f802d5628a8cc32618044b9383211d5a492c56060d427f953b7793a1064029e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c982g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:07:29Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-chmnz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:43Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:43 crc kubenswrapper[4953]: I1203 00:07:43.344933 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-xshk4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bd03a758-0ca9-47d8-86f7-18b2d610df43\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47gfh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47gfh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:07:31Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-xshk4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:43Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:43 crc kubenswrapper[4953]: I1203 00:07:43.347119 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xshk4" Dec 03 00:07:43 crc kubenswrapper[4953]: E1203 00:07:43.347235 4953 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xshk4" podUID="bd03a758-0ca9-47d8-86f7-18b2d610df43" Dec 03 00:07:43 crc kubenswrapper[4953]: I1203 00:07:43.363991 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cf821840-5353-4620-8f73-99b56b371ab0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cae5e2994fcc24dc9b2ea73e79572615b3311edc42642b7f278fb84123989b23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://29e4b9d1b70a93fb133e0759ad148ae133766a66efeae3ff06a21dfae0e805ff\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dec43536772aa7c3f5a662fa7f3bd73b1cae2791ad454f85792287828d0ffeed\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://034e0bac6892d361700b8f33d6427b8a55e2bd3f89b14ea77a374c6be94d25e2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://761973393a93c59c23e69f41b8ac0d16abc35abd61ecbc6d51a42d94cd664233\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T00:07:09Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 00:07:03.713374 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 00:07:03.715052 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1528563048/tls.crt::/tmp/serving-cert-1528563048/tls.key\\\\\\\"\\\\nI1203 00:07:09.300052 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 00:07:09.304192 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 00:07:09.304255 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 00:07:09.304309 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 00:07:09.304323 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 00:07:09.313563 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 00:07:09.313597 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 00:07:09.313603 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 00:07:09.313609 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 00:07:09.313613 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 00:07:09.313618 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 00:07:09.313622 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 00:07:09.313709 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 00:07:09.327347 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T00:06:53Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c143d4232fb8bf7ef5999a664e431a00ab0db030f4d31a28d0bb21e00e38581\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:53Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72af6c27296eae2e275d455795c68f0d272869d63642b8353b17d466d0c901cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://72af6c27296eae2e275d455795c68f0d272869d63642b8353b17d466d0c901cf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:06:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:06:51Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:43Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:43 crc kubenswrapper[4953]: I1203 00:07:43.374680 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:43 crc kubenswrapper[4953]: I1203 00:07:43.374763 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:43 crc kubenswrapper[4953]: I1203 00:07:43.374780 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:43 crc kubenswrapper[4953]: I1203 00:07:43.374804 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:43 crc kubenswrapper[4953]: I1203 00:07:43.374845 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:43Z","lastTransitionTime":"2025-12-03T00:07:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:43 crc kubenswrapper[4953]: I1203 00:07:43.378379 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:10Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:43Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:43 crc kubenswrapper[4953]: E1203 00:07:43.388313 4953 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:07:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:07:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:43Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:07:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:07:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:43Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"eb4ae5a6-0f6f-49ef-a027-1cdee9d3c536\\\",\\\"systemUUID\\\":\\\"4c1f3d44-3d3d-425d-993f-471223206a89\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:43Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:43 crc kubenswrapper[4953]: I1203 00:07:43.391781 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:43 crc kubenswrapper[4953]: I1203 00:07:43.391812 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:43 crc kubenswrapper[4953]: I1203 00:07:43.391823 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:43 crc kubenswrapper[4953]: I1203 00:07:43.391838 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:43 crc kubenswrapper[4953]: I1203 00:07:43.391851 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:43Z","lastTransitionTime":"2025-12-03T00:07:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:43 crc kubenswrapper[4953]: I1203 00:07:43.395436 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce8b414e1fc22eb30decb6a6411b672e7f443055cb89ebb029fbf6ffc1d5dca7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:43Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:43 crc kubenswrapper[4953]: E1203 00:07:43.410412 4953 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:07:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:07:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:43Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:07:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:07:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:43Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"eb4ae5a6-0f6f-49ef-a027-1cdee9d3c536\\\",\\\"systemUUID\\\":\\\"4c1f3d44-3d3d-425d-993f-471223206a89\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:43Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:43 crc kubenswrapper[4953]: I1203 00:07:43.412428 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-6dfbn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53ffb1e9-9a2e-4869-80d7-776f4aa7ef8b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ab4205c0d38ad6a0c32077eadb1fd5fd3a279d9c8b8a2ffbe693ec2c4018370c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ct26c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7df84405eda520ad5a22d35750edf07ef3c3977a756ad5739c54f91cba9da7b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ct26c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:07:16Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-6dfbn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:43Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:43 crc kubenswrapper[4953]: I1203 00:07:43.413863 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:43 crc kubenswrapper[4953]: I1203 00:07:43.413893 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:43 crc kubenswrapper[4953]: I1203 00:07:43.413902 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:43 crc kubenswrapper[4953]: I1203 00:07:43.413935 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:43 crc kubenswrapper[4953]: I1203 00:07:43.413947 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:43Z","lastTransitionTime":"2025-12-03T00:07:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:43 crc kubenswrapper[4953]: E1203 00:07:43.426489 4953 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:07:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:07:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:43Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:07:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:07:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:43Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"eb4ae5a6-0f6f-49ef-a027-1cdee9d3c536\\\",\\\"systemUUID\\\":\\\"4c1f3d44-3d3d-425d-993f-471223206a89\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:43Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:43 crc kubenswrapper[4953]: I1203 00:07:43.429878 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-txffx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8816ebee-c735-4b8d-b1ec-65349707b347\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:16Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:16Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4b85224419a4075acb340e4b11095f2adb5c316e285afa4e19295c2181c0d24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0f6f645a58993f052a6f365bb808f62d99f7fea74886c5cc022b2d86112321a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a61b951a9cbab2c48c0f2df8351f97d7ff71bf3674f641fe5eac30155232b224\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd5f428de229749485267e81d34e31b708728b52474032e88460e54747f21e11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://00a2ea4a9bb870fce30e7fa845f787f184e7a19e42c95511655ea8c5f84a26d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc0a5d4cc9d614d8492626a1b114bdc52f510eb5965d842c02c083f08d4849a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://efbfc74a8598c47a658c06b597fa36d2a5ac8d02265d2864fd2832dd11ecb2cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://efbfc74a8598c47a658c06b597fa36d2a5ac8d02265d2864fd2832dd11ecb2cc\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T00:07:29Z\\\",\\\"message\\\":\\\" 6400 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1203 00:07:29.614806 6400 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1203 00:07:29.614821 6400 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1203 00:07:29.614822 6400 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 00:07:29.614854 6400 factory.go:656] Stopping watch factory\\\\nI1203 00:07:29.614869 6400 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1203 00:07:29.614878 6400 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1203 00:07:29.614875 6400 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1203 00:07:29.614894 6400 handler.go:208] Removed *v1.Node event handler 7\\\\nI1203 00:07:29.614906 6400 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1203 00:07:29.614915 6400 reflector.go:311] Stopping reflector *v1.NetworkAttachmentDefinition (0s) from github.com/k8snetworkplumbingwg/network-attachment-definition-client/pkg/client/informers/externalversions/factory.go:117\\\\nI1203 00:07:29.615144 6400 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 00:07:29.615318 6400 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T00:07:28Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-txffx_openshift-ovn-kubernetes(8816ebee-c735-4b8d-b1ec-65349707b347)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7deadc8e135ff1bc7c5a2f093e489bb5b4b6f166bc874c6acc2a841c5397f11e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af4c7764fca87144d01f3738ea010308744ccf50c54b7c471ccc213112de00d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af4c7764fca87144d01f3738ea010308744ccf50c54b7c471ccc213112de00d5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:07:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:07:16Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-txffx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:43Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:43 crc kubenswrapper[4953]: I1203 00:07:43.430608 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:43 crc kubenswrapper[4953]: I1203 00:07:43.430634 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:43 crc kubenswrapper[4953]: I1203 00:07:43.430643 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:43 crc kubenswrapper[4953]: I1203 00:07:43.430660 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:43 crc kubenswrapper[4953]: I1203 00:07:43.430671 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:43Z","lastTransitionTime":"2025-12-03T00:07:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:43 crc kubenswrapper[4953]: E1203 00:07:43.441701 4953 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:07:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:07:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:43Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:07:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:07:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:43Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"eb4ae5a6-0f6f-49ef-a027-1cdee9d3c536\\\",\\\"systemUUID\\\":\\\"4c1f3d44-3d3d-425d-993f-471223206a89\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:43Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:43 crc kubenswrapper[4953]: I1203 00:07:43.445132 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:43 crc kubenswrapper[4953]: I1203 00:07:43.445165 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:43 crc kubenswrapper[4953]: I1203 00:07:43.445174 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:43 crc kubenswrapper[4953]: I1203 00:07:43.445187 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:43 crc kubenswrapper[4953]: I1203 00:07:43.445198 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:43Z","lastTransitionTime":"2025-12-03T00:07:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:43 crc kubenswrapper[4953]: E1203 00:07:43.458651 4953 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:07:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:07:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:43Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:07:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:07:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:43Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"eb4ae5a6-0f6f-49ef-a027-1cdee9d3c536\\\",\\\"systemUUID\\\":\\\"4c1f3d44-3d3d-425d-993f-471223206a89\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:43Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:43 crc kubenswrapper[4953]: E1203 00:07:43.458799 4953 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 03 00:07:43 crc kubenswrapper[4953]: I1203 00:07:43.460757 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:43 crc kubenswrapper[4953]: I1203 00:07:43.460791 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:43 crc kubenswrapper[4953]: I1203 00:07:43.460801 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:43 crc kubenswrapper[4953]: I1203 00:07:43.460817 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:43 crc kubenswrapper[4953]: I1203 00:07:43.460829 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:43Z","lastTransitionTime":"2025-12-03T00:07:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:43 crc kubenswrapper[4953]: I1203 00:07:43.562834 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:43 crc kubenswrapper[4953]: I1203 00:07:43.562883 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:43 crc kubenswrapper[4953]: I1203 00:07:43.562898 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:43 crc kubenswrapper[4953]: I1203 00:07:43.562919 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:43 crc kubenswrapper[4953]: I1203 00:07:43.562934 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:43Z","lastTransitionTime":"2025-12-03T00:07:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:43 crc kubenswrapper[4953]: I1203 00:07:43.665083 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:43 crc kubenswrapper[4953]: I1203 00:07:43.665124 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:43 crc kubenswrapper[4953]: I1203 00:07:43.665139 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:43 crc kubenswrapper[4953]: I1203 00:07:43.665160 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:43 crc kubenswrapper[4953]: I1203 00:07:43.665175 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:43Z","lastTransitionTime":"2025-12-03T00:07:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:43 crc kubenswrapper[4953]: I1203 00:07:43.731161 4953 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-txffx_8816ebee-c735-4b8d-b1ec-65349707b347/ovnkube-controller/1.log" Dec 03 00:07:43 crc kubenswrapper[4953]: I1203 00:07:43.733540 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-txffx" event={"ID":"8816ebee-c735-4b8d-b1ec-65349707b347","Type":"ContainerStarted","Data":"dcb805f9ce285d3841e93d588fe7f3d42232382366d42b75bcfb129aefa91c1f"} Dec 03 00:07:43 crc kubenswrapper[4953]: I1203 00:07:43.734065 4953 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-txffx" Dec 03 00:07:43 crc kubenswrapper[4953]: I1203 00:07:43.748564 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:10Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:43Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:43 crc kubenswrapper[4953]: I1203 00:07:43.762326 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-675qk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"43a4bad8-71f4-441a-8302-aeb526a986af\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5b5794fef6477aab5deed8f642e8cc50b37eb1fd0cc6f19c003f34ad31c1e3c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gdrw4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:07:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-675qk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:43Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:43 crc kubenswrapper[4953]: I1203 00:07:43.766897 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:43 crc kubenswrapper[4953]: I1203 00:07:43.766920 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:43 crc kubenswrapper[4953]: I1203 00:07:43.766929 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:43 crc kubenswrapper[4953]: I1203 00:07:43.766956 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:43 crc kubenswrapper[4953]: I1203 00:07:43.766966 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:43Z","lastTransitionTime":"2025-12-03T00:07:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:43 crc kubenswrapper[4953]: I1203 00:07:43.776927 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f94f8c51-c665-4ca3-bf17-7a04ea5c7ade\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72c5cd7dc1c3a8adfedab6ba43a59d80b3efe65f2bb56be95e76cc86a409dcf3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2810ffc1b141fa7da811406053db9d72c708edf5f2bb1dd75e4d52ab35fd24c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c67b4277b20a9f1287f50ef82d0785c9d02543a8e7ed26e0413eae332b866db7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://258d78e1ad9d3dcda6976909ae3fe71aecead5b995fc66b3fd6e2722d80b5ab8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:06:51Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:43Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:43 crc kubenswrapper[4953]: I1203 00:07:43.791694 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5652d0f3-3fd7-4af6-8941-9119e1177641\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2030aef5b7170991268fa5dc7e7ed8ce1fbe3a2765165a10c25251426575330f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ab6052f1abf17225155ce1bd86afa742e6e78f507b7deedd7161cd8fba9120b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://650b6c79cda2e1768ba1d8763db799e210ce2748893512068d25e60f821937ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d1d60162f7f6885d4f51b4ab44163cfae6d2d04355b1dd710e26b022450005df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d1d60162f7f6885d4f51b4ab44163cfae6d2d04355b1dd710e26b022450005df\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:06:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:06:51Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:43Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:43 crc kubenswrapper[4953]: I1203 00:07:43.807908 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09657a9f448a4c92a759c73a0a892076d7eb46393e6fd565e5f73f34c71aac23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5af5eac4a58870f63497f99b503cfa5fb922278095e0332ec50376925ecf2109\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:43Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:43 crc kubenswrapper[4953]: I1203 00:07:43.831830 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-wj6rz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"96aa67a0-15dd-420d-a1d5-ed239d7df1fb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01815cfd3f58c3f3f5f2a894c1790e0b69c5060c48dc0b1deff050b6ad7af916\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kf62m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b12bfb5bdf9fd4f2c50a335fc82df6ee0440eaf8741897da03aa2dae368cc6b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b12bfb5bdf9fd4f2c50a335fc82df6ee0440eaf8741897da03aa2dae368cc6b4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:07:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kf62m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83aaa5fa28110bc0ef05aff75cd1442bc9bbdd33b822ae2ef22d4e4ab83238bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://83aaa5fa28110bc0ef05aff75cd1442bc9bbdd33b822ae2ef22d4e4ab83238bc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:07:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:07:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kf62m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ab6213a33a19fedebce6d74a0ebbb3b3a2de5687504e21cc24093cc42ade635\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ab6213a33a19fedebce6d74a0ebbb3b3a2de5687504e21cc24093cc42ade635\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:07:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:07:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kf62m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://de480488843afa6b8dede1e1b42b5181dca1993e2f39dc1b461a18a57e3668e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://de480488843afa6b8dede1e1b42b5181dca1993e2f39dc1b461a18a57e3668e1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:07:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kf62m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://97be4ba692bc61078d4b0ba39fb0c12772548c6c9251ff3356ea702f151223cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://97be4ba692bc61078d4b0ba39fb0c12772548c6c9251ff3356ea702f151223cd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:07:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kf62m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1a1f009fcbd8af9217db1dc97f662719f8c1db232df0e2804e44c881abb1df2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b1a1f009fcbd8af9217db1dc97f662719f8c1db232df0e2804e44c881abb1df2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:07:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:07:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kf62m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:07:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-wj6rz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:43Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:43 crc kubenswrapper[4953]: I1203 00:07:43.844065 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-kl5s9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d85e5c97-16ad-4bd7-9e73-b7f12678be97\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c21df91a9e4e3165686882b2eb559498224063dc633c3a63f958598108eca29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j24lb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:07:16Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-kl5s9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:43Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:43 crc kubenswrapper[4953]: I1203 00:07:43.855599 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-xshk4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bd03a758-0ca9-47d8-86f7-18b2d610df43\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47gfh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47gfh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:07:31Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-xshk4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:43Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:43 crc kubenswrapper[4953]: I1203 00:07:43.869874 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:43 crc kubenswrapper[4953]: I1203 00:07:43.869909 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:43 crc kubenswrapper[4953]: I1203 00:07:43.869920 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:43 crc kubenswrapper[4953]: I1203 00:07:43.869936 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:43 crc kubenswrapper[4953]: I1203 00:07:43.869948 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:43Z","lastTransitionTime":"2025-12-03T00:07:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:43 crc kubenswrapper[4953]: I1203 00:07:43.875672 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cf821840-5353-4620-8f73-99b56b371ab0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cae5e2994fcc24dc9b2ea73e79572615b3311edc42642b7f278fb84123989b23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://29e4b9d1b70a93fb133e0759ad148ae133766a66efeae3ff06a21dfae0e805ff\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dec43536772aa7c3f5a662fa7f3bd73b1cae2791ad454f85792287828d0ffeed\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://034e0bac6892d361700b8f33d6427b8a55e2bd3f89b14ea77a374c6be94d25e2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://761973393a93c59c23e69f41b8ac0d16abc35abd61ecbc6d51a42d94cd664233\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T00:07:09Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 00:07:03.713374 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 00:07:03.715052 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1528563048/tls.crt::/tmp/serving-cert-1528563048/tls.key\\\\\\\"\\\\nI1203 00:07:09.300052 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 00:07:09.304192 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 00:07:09.304255 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 00:07:09.304309 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 00:07:09.304323 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 00:07:09.313563 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 00:07:09.313597 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 00:07:09.313603 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 00:07:09.313609 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 00:07:09.313613 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 00:07:09.313618 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 00:07:09.313622 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 00:07:09.313709 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 00:07:09.327347 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T00:06:53Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c143d4232fb8bf7ef5999a664e431a00ab0db030f4d31a28d0bb21e00e38581\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:53Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72af6c27296eae2e275d455795c68f0d272869d63642b8353b17d466d0c901cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://72af6c27296eae2e275d455795c68f0d272869d63642b8353b17d466d0c901cf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:06:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:06:51Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:43Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:43 crc kubenswrapper[4953]: I1203 00:07:43.891718 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:10Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:43Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:43 crc kubenswrapper[4953]: I1203 00:07:43.911178 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce8b414e1fc22eb30decb6a6411b672e7f443055cb89ebb029fbf6ffc1d5dca7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:43Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:43 crc kubenswrapper[4953]: I1203 00:07:43.927947 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-6dfbn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53ffb1e9-9a2e-4869-80d7-776f4aa7ef8b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ab4205c0d38ad6a0c32077eadb1fd5fd3a279d9c8b8a2ffbe693ec2c4018370c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ct26c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7df84405eda520ad5a22d35750edf07ef3c3977a756ad5739c54f91cba9da7b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ct26c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:07:16Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-6dfbn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:43Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:43 crc kubenswrapper[4953]: I1203 00:07:43.962277 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-txffx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8816ebee-c735-4b8d-b1ec-65349707b347\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:16Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:16Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4b85224419a4075acb340e4b11095f2adb5c316e285afa4e19295c2181c0d24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0f6f645a58993f052a6f365bb808f62d99f7fea74886c5cc022b2d86112321a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a61b951a9cbab2c48c0f2df8351f97d7ff71bf3674f641fe5eac30155232b224\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd5f428de229749485267e81d34e31b708728b52474032e88460e54747f21e11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://00a2ea4a9bb870fce30e7fa845f787f184e7a19e42c95511655ea8c5f84a26d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc0a5d4cc9d614d8492626a1b114bdc52f510eb5965d842c02c083f08d4849a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dcb805f9ce285d3841e93d588fe7f3d42232382366d42b75bcfb129aefa91c1f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://efbfc74a8598c47a658c06b597fa36d2a5ac8d02265d2864fd2832dd11ecb2cc\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T00:07:29Z\\\",\\\"message\\\":\\\" 6400 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1203 00:07:29.614806 6400 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1203 00:07:29.614821 6400 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1203 00:07:29.614822 6400 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 00:07:29.614854 6400 factory.go:656] Stopping watch factory\\\\nI1203 00:07:29.614869 6400 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1203 00:07:29.614878 6400 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1203 00:07:29.614875 6400 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1203 00:07:29.614894 6400 handler.go:208] Removed *v1.Node event handler 7\\\\nI1203 00:07:29.614906 6400 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1203 00:07:29.614915 6400 reflector.go:311] Stopping reflector *v1.NetworkAttachmentDefinition (0s) from github.com/k8snetworkplumbingwg/network-attachment-definition-client/pkg/client/informers/externalversions/factory.go:117\\\\nI1203 00:07:29.615144 6400 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 00:07:29.615318 6400 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T00:07:28Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7deadc8e135ff1bc7c5a2f093e489bb5b4b6f166bc874c6acc2a841c5397f11e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af4c7764fca87144d01f3738ea010308744ccf50c54b7c471ccc213112de00d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af4c7764fca87144d01f3738ea010308744ccf50c54b7c471ccc213112de00d5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:07:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:07:16Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-txffx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:43Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:43 crc kubenswrapper[4953]: I1203 00:07:43.972765 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:43 crc kubenswrapper[4953]: I1203 00:07:43.972826 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:43 crc kubenswrapper[4953]: I1203 00:07:43.972846 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:43 crc kubenswrapper[4953]: I1203 00:07:43.972871 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:43 crc kubenswrapper[4953]: I1203 00:07:43.972889 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:43Z","lastTransitionTime":"2025-12-03T00:07:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:43 crc kubenswrapper[4953]: I1203 00:07:43.973345 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5cmlc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7e11a1d3-ecee-4bf0-87bf-085fab97b73d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d42390b9c53efc63add8ddcebf03c976f5e5448eaa2c34e076b2dc90dbac60fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6x2sc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:07:18Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5cmlc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:43Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:43 crc kubenswrapper[4953]: I1203 00:07:43.990635 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-chmnz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a12fe6d-d25a-4b08-90da-cc9e510980a7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://711ab839cb9c0a90554dadbd45d810fc78aac09739340482fef8de8743b77283\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c982g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f802d5628a8cc32618044b9383211d5a492c56060d427f953b7793a1064029e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c982g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:07:29Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-chmnz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:43Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:44 crc kubenswrapper[4953]: I1203 00:07:44.005179 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:44Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:44 crc kubenswrapper[4953]: I1203 00:07:44.015605 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2ed9c7fa96c8951fbc1728fe7a91fefdd999077285621df9c1abedf9ae0c449\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:44Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:44 crc kubenswrapper[4953]: I1203 00:07:44.075421 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:44 crc kubenswrapper[4953]: I1203 00:07:44.075544 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:44 crc kubenswrapper[4953]: I1203 00:07:44.075564 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:44 crc kubenswrapper[4953]: I1203 00:07:44.075586 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:44 crc kubenswrapper[4953]: I1203 00:07:44.075602 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:44Z","lastTransitionTime":"2025-12-03T00:07:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:44 crc kubenswrapper[4953]: I1203 00:07:44.178620 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:44 crc kubenswrapper[4953]: I1203 00:07:44.178667 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:44 crc kubenswrapper[4953]: I1203 00:07:44.178684 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:44 crc kubenswrapper[4953]: I1203 00:07:44.178706 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:44 crc kubenswrapper[4953]: I1203 00:07:44.178723 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:44Z","lastTransitionTime":"2025-12-03T00:07:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:44 crc kubenswrapper[4953]: I1203 00:07:44.281774 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:44 crc kubenswrapper[4953]: I1203 00:07:44.281848 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:44 crc kubenswrapper[4953]: I1203 00:07:44.281872 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:44 crc kubenswrapper[4953]: I1203 00:07:44.281903 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:44 crc kubenswrapper[4953]: I1203 00:07:44.281957 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:44Z","lastTransitionTime":"2025-12-03T00:07:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:44 crc kubenswrapper[4953]: I1203 00:07:44.348035 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 00:07:44 crc kubenswrapper[4953]: I1203 00:07:44.348103 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 00:07:44 crc kubenswrapper[4953]: E1203 00:07:44.348185 4953 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 00:07:44 crc kubenswrapper[4953]: I1203 00:07:44.348201 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 00:07:44 crc kubenswrapper[4953]: E1203 00:07:44.348353 4953 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 00:07:44 crc kubenswrapper[4953]: E1203 00:07:44.348451 4953 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 00:07:44 crc kubenswrapper[4953]: I1203 00:07:44.385135 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:44 crc kubenswrapper[4953]: I1203 00:07:44.385200 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:44 crc kubenswrapper[4953]: I1203 00:07:44.385216 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:44 crc kubenswrapper[4953]: I1203 00:07:44.385240 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:44 crc kubenswrapper[4953]: I1203 00:07:44.385257 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:44Z","lastTransitionTime":"2025-12-03T00:07:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:44 crc kubenswrapper[4953]: I1203 00:07:44.488778 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:44 crc kubenswrapper[4953]: I1203 00:07:44.488855 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:44 crc kubenswrapper[4953]: I1203 00:07:44.488876 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:44 crc kubenswrapper[4953]: I1203 00:07:44.488909 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:44 crc kubenswrapper[4953]: I1203 00:07:44.488929 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:44Z","lastTransitionTime":"2025-12-03T00:07:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:44 crc kubenswrapper[4953]: I1203 00:07:44.591934 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:44 crc kubenswrapper[4953]: I1203 00:07:44.592026 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:44 crc kubenswrapper[4953]: I1203 00:07:44.592044 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:44 crc kubenswrapper[4953]: I1203 00:07:44.592069 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:44 crc kubenswrapper[4953]: I1203 00:07:44.592085 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:44Z","lastTransitionTime":"2025-12-03T00:07:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:44 crc kubenswrapper[4953]: I1203 00:07:44.695206 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:44 crc kubenswrapper[4953]: I1203 00:07:44.695261 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:44 crc kubenswrapper[4953]: I1203 00:07:44.695278 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:44 crc kubenswrapper[4953]: I1203 00:07:44.695302 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:44 crc kubenswrapper[4953]: I1203 00:07:44.695319 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:44Z","lastTransitionTime":"2025-12-03T00:07:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:44 crc kubenswrapper[4953]: I1203 00:07:44.740228 4953 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-txffx_8816ebee-c735-4b8d-b1ec-65349707b347/ovnkube-controller/2.log" Dec 03 00:07:44 crc kubenswrapper[4953]: I1203 00:07:44.741484 4953 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-txffx_8816ebee-c735-4b8d-b1ec-65349707b347/ovnkube-controller/1.log" Dec 03 00:07:44 crc kubenswrapper[4953]: I1203 00:07:44.746725 4953 generic.go:334] "Generic (PLEG): container finished" podID="8816ebee-c735-4b8d-b1ec-65349707b347" containerID="dcb805f9ce285d3841e93d588fe7f3d42232382366d42b75bcfb129aefa91c1f" exitCode=1 Dec 03 00:07:44 crc kubenswrapper[4953]: I1203 00:07:44.746788 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-txffx" event={"ID":"8816ebee-c735-4b8d-b1ec-65349707b347","Type":"ContainerDied","Data":"dcb805f9ce285d3841e93d588fe7f3d42232382366d42b75bcfb129aefa91c1f"} Dec 03 00:07:44 crc kubenswrapper[4953]: I1203 00:07:44.746857 4953 scope.go:117] "RemoveContainer" containerID="efbfc74a8598c47a658c06b597fa36d2a5ac8d02265d2864fd2832dd11ecb2cc" Dec 03 00:07:44 crc kubenswrapper[4953]: I1203 00:07:44.748533 4953 scope.go:117] "RemoveContainer" containerID="dcb805f9ce285d3841e93d588fe7f3d42232382366d42b75bcfb129aefa91c1f" Dec 03 00:07:44 crc kubenswrapper[4953]: E1203 00:07:44.751226 4953 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-txffx_openshift-ovn-kubernetes(8816ebee-c735-4b8d-b1ec-65349707b347)\"" pod="openshift-ovn-kubernetes/ovnkube-node-txffx" podUID="8816ebee-c735-4b8d-b1ec-65349707b347" Dec 03 00:07:44 crc kubenswrapper[4953]: I1203 00:07:44.768050 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2ed9c7fa96c8951fbc1728fe7a91fefdd999077285621df9c1abedf9ae0c449\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:44Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:44 crc kubenswrapper[4953]: I1203 00:07:44.786578 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:44Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:44 crc kubenswrapper[4953]: I1203 00:07:44.799356 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:44 crc kubenswrapper[4953]: I1203 00:07:44.799410 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:44 crc kubenswrapper[4953]: I1203 00:07:44.799428 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:44 crc kubenswrapper[4953]: I1203 00:07:44.799453 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:44 crc kubenswrapper[4953]: I1203 00:07:44.799470 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:44Z","lastTransitionTime":"2025-12-03T00:07:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:44 crc kubenswrapper[4953]: I1203 00:07:44.805777 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:10Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:44Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:44 crc kubenswrapper[4953]: I1203 00:07:44.826249 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-675qk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"43a4bad8-71f4-441a-8302-aeb526a986af\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5b5794fef6477aab5deed8f642e8cc50b37eb1fd0cc6f19c003f34ad31c1e3c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gdrw4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:07:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-675qk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:44Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:44 crc kubenswrapper[4953]: I1203 00:07:44.850290 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-wj6rz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"96aa67a0-15dd-420d-a1d5-ed239d7df1fb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01815cfd3f58c3f3f5f2a894c1790e0b69c5060c48dc0b1deff050b6ad7af916\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kf62m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b12bfb5bdf9fd4f2c50a335fc82df6ee0440eaf8741897da03aa2dae368cc6b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b12bfb5bdf9fd4f2c50a335fc82df6ee0440eaf8741897da03aa2dae368cc6b4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:07:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kf62m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83aaa5fa28110bc0ef05aff75cd1442bc9bbdd33b822ae2ef22d4e4ab83238bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://83aaa5fa28110bc0ef05aff75cd1442bc9bbdd33b822ae2ef22d4e4ab83238bc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:07:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:07:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kf62m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ab6213a33a19fedebce6d74a0ebbb3b3a2de5687504e21cc24093cc42ade635\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ab6213a33a19fedebce6d74a0ebbb3b3a2de5687504e21cc24093cc42ade635\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:07:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:07:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kf62m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://de480488843afa6b8dede1e1b42b5181dca1993e2f39dc1b461a18a57e3668e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://de480488843afa6b8dede1e1b42b5181dca1993e2f39dc1b461a18a57e3668e1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:07:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kf62m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://97be4ba692bc61078d4b0ba39fb0c12772548c6c9251ff3356ea702f151223cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://97be4ba692bc61078d4b0ba39fb0c12772548c6c9251ff3356ea702f151223cd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:07:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kf62m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1a1f009fcbd8af9217db1dc97f662719f8c1db232df0e2804e44c881abb1df2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b1a1f009fcbd8af9217db1dc97f662719f8c1db232df0e2804e44c881abb1df2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:07:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:07:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kf62m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:07:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-wj6rz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:44Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:44 crc kubenswrapper[4953]: I1203 00:07:44.862316 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-kl5s9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d85e5c97-16ad-4bd7-9e73-b7f12678be97\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c21df91a9e4e3165686882b2eb559498224063dc633c3a63f958598108eca29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j24lb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:07:16Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-kl5s9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:44Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:44 crc kubenswrapper[4953]: I1203 00:07:44.885199 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f94f8c51-c665-4ca3-bf17-7a04ea5c7ade\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72c5cd7dc1c3a8adfedab6ba43a59d80b3efe65f2bb56be95e76cc86a409dcf3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2810ffc1b141fa7da811406053db9d72c708edf5f2bb1dd75e4d52ab35fd24c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c67b4277b20a9f1287f50ef82d0785c9d02543a8e7ed26e0413eae332b866db7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://258d78e1ad9d3dcda6976909ae3fe71aecead5b995fc66b3fd6e2722d80b5ab8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:06:51Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:44Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:44 crc kubenswrapper[4953]: I1203 00:07:44.902787 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:44 crc kubenswrapper[4953]: I1203 00:07:44.902856 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:44 crc kubenswrapper[4953]: I1203 00:07:44.902879 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:44 crc kubenswrapper[4953]: I1203 00:07:44.902911 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:44 crc kubenswrapper[4953]: I1203 00:07:44.902930 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:44Z","lastTransitionTime":"2025-12-03T00:07:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:44 crc kubenswrapper[4953]: I1203 00:07:44.904648 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5652d0f3-3fd7-4af6-8941-9119e1177641\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2030aef5b7170991268fa5dc7e7ed8ce1fbe3a2765165a10c25251426575330f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ab6052f1abf17225155ce1bd86afa742e6e78f507b7deedd7161cd8fba9120b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://650b6c79cda2e1768ba1d8763db799e210ce2748893512068d25e60f821937ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d1d60162f7f6885d4f51b4ab44163cfae6d2d04355b1dd710e26b022450005df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d1d60162f7f6885d4f51b4ab44163cfae6d2d04355b1dd710e26b022450005df\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:06:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:06:51Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:44Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:44 crc kubenswrapper[4953]: I1203 00:07:44.925775 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09657a9f448a4c92a759c73a0a892076d7eb46393e6fd565e5f73f34c71aac23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5af5eac4a58870f63497f99b503cfa5fb922278095e0332ec50376925ecf2109\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:44Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:44 crc kubenswrapper[4953]: I1203 00:07:44.944044 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-6dfbn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53ffb1e9-9a2e-4869-80d7-776f4aa7ef8b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ab4205c0d38ad6a0c32077eadb1fd5fd3a279d9c8b8a2ffbe693ec2c4018370c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ct26c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7df84405eda520ad5a22d35750edf07ef3c3977a756ad5739c54f91cba9da7b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ct26c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:07:16Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-6dfbn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:44Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:44 crc kubenswrapper[4953]: I1203 00:07:44.975142 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-txffx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8816ebee-c735-4b8d-b1ec-65349707b347\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:16Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:16Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4b85224419a4075acb340e4b11095f2adb5c316e285afa4e19295c2181c0d24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0f6f645a58993f052a6f365bb808f62d99f7fea74886c5cc022b2d86112321a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a61b951a9cbab2c48c0f2df8351f97d7ff71bf3674f641fe5eac30155232b224\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd5f428de229749485267e81d34e31b708728b52474032e88460e54747f21e11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://00a2ea4a9bb870fce30e7fa845f787f184e7a19e42c95511655ea8c5f84a26d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc0a5d4cc9d614d8492626a1b114bdc52f510eb5965d842c02c083f08d4849a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dcb805f9ce285d3841e93d588fe7f3d42232382366d42b75bcfb129aefa91c1f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://efbfc74a8598c47a658c06b597fa36d2a5ac8d02265d2864fd2832dd11ecb2cc\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T00:07:29Z\\\",\\\"message\\\":\\\" 6400 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1203 00:07:29.614806 6400 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1203 00:07:29.614821 6400 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1203 00:07:29.614822 6400 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 00:07:29.614854 6400 factory.go:656] Stopping watch factory\\\\nI1203 00:07:29.614869 6400 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1203 00:07:29.614878 6400 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1203 00:07:29.614875 6400 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1203 00:07:29.614894 6400 handler.go:208] Removed *v1.Node event handler 7\\\\nI1203 00:07:29.614906 6400 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1203 00:07:29.614915 6400 reflector.go:311] Stopping reflector *v1.NetworkAttachmentDefinition (0s) from github.com/k8snetworkplumbingwg/network-attachment-definition-client/pkg/client/informers/externalversions/factory.go:117\\\\nI1203 00:07:29.615144 6400 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 00:07:29.615318 6400 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T00:07:28Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dcb805f9ce285d3841e93d588fe7f3d42232382366d42b75bcfb129aefa91c1f\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T00:07:44Z\\\",\\\"message\\\":\\\"mers/externalversions/factory.go:140\\\\nI1203 00:07:44.043157 6594 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1203 00:07:44.043232 6594 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 00:07:44.043450 6594 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 00:07:44.043811 6594 reflector.go:311] Stopping reflector *v1.NetworkAttachmentDefinition (0s) from github.com/k8snetworkplumbingwg/network-attachment-definition-client/pkg/client/informers/externalversions/factory.go:117\\\\nI1203 00:07:44.044028 6594 reflector.go:311] Stopping reflector *v1.AdminPolicyBasedExternalRoute (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140\\\\nI1203 00:07:44.044197 6594 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI1203 00:07:44.044257 6594 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T00:07:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7deadc8e135ff1bc7c5a2f093e489bb5b4b6f166bc874c6acc2a841c5397f11e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af4c7764fca87144d01f3738ea010308744ccf50c54b7c471ccc213112de00d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af4c7764fca87144d01f3738ea010308744ccf50c54b7c471ccc213112de00d5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:07:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:07:16Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-txffx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:44Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:44 crc kubenswrapper[4953]: I1203 00:07:44.991321 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5cmlc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7e11a1d3-ecee-4bf0-87bf-085fab97b73d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d42390b9c53efc63add8ddcebf03c976f5e5448eaa2c34e076b2dc90dbac60fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6x2sc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:07:18Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5cmlc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:44Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:45 crc kubenswrapper[4953]: I1203 00:07:45.006210 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:45 crc kubenswrapper[4953]: I1203 00:07:45.006270 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:45 crc kubenswrapper[4953]: I1203 00:07:45.006296 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:45 crc kubenswrapper[4953]: I1203 00:07:45.006326 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:45 crc kubenswrapper[4953]: I1203 00:07:45.006349 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:45Z","lastTransitionTime":"2025-12-03T00:07:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:45 crc kubenswrapper[4953]: I1203 00:07:45.011396 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-chmnz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a12fe6d-d25a-4b08-90da-cc9e510980a7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://711ab839cb9c0a90554dadbd45d810fc78aac09739340482fef8de8743b77283\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c982g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f802d5628a8cc32618044b9383211d5a492c56060d427f953b7793a1064029e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c982g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:07:29Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-chmnz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:45Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:45 crc kubenswrapper[4953]: I1203 00:07:45.027228 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-xshk4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bd03a758-0ca9-47d8-86f7-18b2d610df43\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47gfh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47gfh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:07:31Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-xshk4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:45Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:45 crc kubenswrapper[4953]: I1203 00:07:45.048970 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cf821840-5353-4620-8f73-99b56b371ab0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cae5e2994fcc24dc9b2ea73e79572615b3311edc42642b7f278fb84123989b23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://29e4b9d1b70a93fb133e0759ad148ae133766a66efeae3ff06a21dfae0e805ff\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dec43536772aa7c3f5a662fa7f3bd73b1cae2791ad454f85792287828d0ffeed\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://034e0bac6892d361700b8f33d6427b8a55e2bd3f89b14ea77a374c6be94d25e2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://761973393a93c59c23e69f41b8ac0d16abc35abd61ecbc6d51a42d94cd664233\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T00:07:09Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 00:07:03.713374 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 00:07:03.715052 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1528563048/tls.crt::/tmp/serving-cert-1528563048/tls.key\\\\\\\"\\\\nI1203 00:07:09.300052 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 00:07:09.304192 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 00:07:09.304255 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 00:07:09.304309 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 00:07:09.304323 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 00:07:09.313563 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 00:07:09.313597 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 00:07:09.313603 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 00:07:09.313609 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 00:07:09.313613 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 00:07:09.313618 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 00:07:09.313622 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 00:07:09.313709 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 00:07:09.327347 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T00:06:53Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c143d4232fb8bf7ef5999a664e431a00ab0db030f4d31a28d0bb21e00e38581\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:53Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72af6c27296eae2e275d455795c68f0d272869d63642b8353b17d466d0c901cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://72af6c27296eae2e275d455795c68f0d272869d63642b8353b17d466d0c901cf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:06:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:06:51Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:45Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:45 crc kubenswrapper[4953]: I1203 00:07:45.063871 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:10Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:45Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:45 crc kubenswrapper[4953]: I1203 00:07:45.081746 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce8b414e1fc22eb30decb6a6411b672e7f443055cb89ebb029fbf6ffc1d5dca7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:45Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:45 crc kubenswrapper[4953]: I1203 00:07:45.113568 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:45 crc kubenswrapper[4953]: I1203 00:07:45.113621 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:45 crc kubenswrapper[4953]: I1203 00:07:45.113634 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:45 crc kubenswrapper[4953]: I1203 00:07:45.113652 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:45 crc kubenswrapper[4953]: I1203 00:07:45.113665 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:45Z","lastTransitionTime":"2025-12-03T00:07:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:45 crc kubenswrapper[4953]: I1203 00:07:45.216298 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:45 crc kubenswrapper[4953]: I1203 00:07:45.216374 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:45 crc kubenswrapper[4953]: I1203 00:07:45.216397 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:45 crc kubenswrapper[4953]: I1203 00:07:45.216425 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:45 crc kubenswrapper[4953]: I1203 00:07:45.216442 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:45Z","lastTransitionTime":"2025-12-03T00:07:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:45 crc kubenswrapper[4953]: I1203 00:07:45.319463 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:45 crc kubenswrapper[4953]: I1203 00:07:45.319518 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:45 crc kubenswrapper[4953]: I1203 00:07:45.319536 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:45 crc kubenswrapper[4953]: I1203 00:07:45.319561 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:45 crc kubenswrapper[4953]: I1203 00:07:45.319580 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:45Z","lastTransitionTime":"2025-12-03T00:07:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:45 crc kubenswrapper[4953]: I1203 00:07:45.347324 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xshk4" Dec 03 00:07:45 crc kubenswrapper[4953]: E1203 00:07:45.347532 4953 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xshk4" podUID="bd03a758-0ca9-47d8-86f7-18b2d610df43" Dec 03 00:07:45 crc kubenswrapper[4953]: I1203 00:07:45.422172 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:45 crc kubenswrapper[4953]: I1203 00:07:45.422237 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:45 crc kubenswrapper[4953]: I1203 00:07:45.422258 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:45 crc kubenswrapper[4953]: I1203 00:07:45.422286 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:45 crc kubenswrapper[4953]: I1203 00:07:45.422304 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:45Z","lastTransitionTime":"2025-12-03T00:07:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:45 crc kubenswrapper[4953]: I1203 00:07:45.525294 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:45 crc kubenswrapper[4953]: I1203 00:07:45.525426 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:45 crc kubenswrapper[4953]: I1203 00:07:45.525452 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:45 crc kubenswrapper[4953]: I1203 00:07:45.525480 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:45 crc kubenswrapper[4953]: I1203 00:07:45.525503 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:45Z","lastTransitionTime":"2025-12-03T00:07:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:45 crc kubenswrapper[4953]: I1203 00:07:45.628158 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:45 crc kubenswrapper[4953]: I1203 00:07:45.628617 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:45 crc kubenswrapper[4953]: I1203 00:07:45.628807 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:45 crc kubenswrapper[4953]: I1203 00:07:45.629033 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:45 crc kubenswrapper[4953]: I1203 00:07:45.629251 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:45Z","lastTransitionTime":"2025-12-03T00:07:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:45 crc kubenswrapper[4953]: I1203 00:07:45.732409 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:45 crc kubenswrapper[4953]: I1203 00:07:45.732466 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:45 crc kubenswrapper[4953]: I1203 00:07:45.732484 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:45 crc kubenswrapper[4953]: I1203 00:07:45.732510 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:45 crc kubenswrapper[4953]: I1203 00:07:45.732528 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:45Z","lastTransitionTime":"2025-12-03T00:07:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:45 crc kubenswrapper[4953]: I1203 00:07:45.753084 4953 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-txffx_8816ebee-c735-4b8d-b1ec-65349707b347/ovnkube-controller/2.log" Dec 03 00:07:45 crc kubenswrapper[4953]: I1203 00:07:45.759295 4953 scope.go:117] "RemoveContainer" containerID="dcb805f9ce285d3841e93d588fe7f3d42232382366d42b75bcfb129aefa91c1f" Dec 03 00:07:45 crc kubenswrapper[4953]: E1203 00:07:45.759629 4953 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-txffx_openshift-ovn-kubernetes(8816ebee-c735-4b8d-b1ec-65349707b347)\"" pod="openshift-ovn-kubernetes/ovnkube-node-txffx" podUID="8816ebee-c735-4b8d-b1ec-65349707b347" Dec 03 00:07:45 crc kubenswrapper[4953]: I1203 00:07:45.779745 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f94f8c51-c665-4ca3-bf17-7a04ea5c7ade\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72c5cd7dc1c3a8adfedab6ba43a59d80b3efe65f2bb56be95e76cc86a409dcf3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2810ffc1b141fa7da811406053db9d72c708edf5f2bb1dd75e4d52ab35fd24c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c67b4277b20a9f1287f50ef82d0785c9d02543a8e7ed26e0413eae332b866db7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://258d78e1ad9d3dcda6976909ae3fe71aecead5b995fc66b3fd6e2722d80b5ab8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:06:51Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:45Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:45 crc kubenswrapper[4953]: I1203 00:07:45.797754 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5652d0f3-3fd7-4af6-8941-9119e1177641\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2030aef5b7170991268fa5dc7e7ed8ce1fbe3a2765165a10c25251426575330f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ab6052f1abf17225155ce1bd86afa742e6e78f507b7deedd7161cd8fba9120b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://650b6c79cda2e1768ba1d8763db799e210ce2748893512068d25e60f821937ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d1d60162f7f6885d4f51b4ab44163cfae6d2d04355b1dd710e26b022450005df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d1d60162f7f6885d4f51b4ab44163cfae6d2d04355b1dd710e26b022450005df\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:06:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:06:51Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:45Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:45 crc kubenswrapper[4953]: I1203 00:07:45.818563 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09657a9f448a4c92a759c73a0a892076d7eb46393e6fd565e5f73f34c71aac23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5af5eac4a58870f63497f99b503cfa5fb922278095e0332ec50376925ecf2109\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:45Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:45 crc kubenswrapper[4953]: I1203 00:07:45.836143 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:45 crc kubenswrapper[4953]: I1203 00:07:45.836205 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:45 crc kubenswrapper[4953]: I1203 00:07:45.836223 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:45 crc kubenswrapper[4953]: I1203 00:07:45.836248 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:45 crc kubenswrapper[4953]: I1203 00:07:45.836265 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:45Z","lastTransitionTime":"2025-12-03T00:07:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:45 crc kubenswrapper[4953]: I1203 00:07:45.843542 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-wj6rz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"96aa67a0-15dd-420d-a1d5-ed239d7df1fb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01815cfd3f58c3f3f5f2a894c1790e0b69c5060c48dc0b1deff050b6ad7af916\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kf62m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b12bfb5bdf9fd4f2c50a335fc82df6ee0440eaf8741897da03aa2dae368cc6b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b12bfb5bdf9fd4f2c50a335fc82df6ee0440eaf8741897da03aa2dae368cc6b4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:07:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kf62m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83aaa5fa28110bc0ef05aff75cd1442bc9bbdd33b822ae2ef22d4e4ab83238bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://83aaa5fa28110bc0ef05aff75cd1442bc9bbdd33b822ae2ef22d4e4ab83238bc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:07:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:07:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kf62m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ab6213a33a19fedebce6d74a0ebbb3b3a2de5687504e21cc24093cc42ade635\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ab6213a33a19fedebce6d74a0ebbb3b3a2de5687504e21cc24093cc42ade635\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:07:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:07:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kf62m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://de480488843afa6b8dede1e1b42b5181dca1993e2f39dc1b461a18a57e3668e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://de480488843afa6b8dede1e1b42b5181dca1993e2f39dc1b461a18a57e3668e1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:07:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kf62m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://97be4ba692bc61078d4b0ba39fb0c12772548c6c9251ff3356ea702f151223cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://97be4ba692bc61078d4b0ba39fb0c12772548c6c9251ff3356ea702f151223cd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:07:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kf62m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1a1f009fcbd8af9217db1dc97f662719f8c1db232df0e2804e44c881abb1df2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b1a1f009fcbd8af9217db1dc97f662719f8c1db232df0e2804e44c881abb1df2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:07:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:07:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kf62m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:07:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-wj6rz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:45Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:45 crc kubenswrapper[4953]: I1203 00:07:45.859275 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-kl5s9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d85e5c97-16ad-4bd7-9e73-b7f12678be97\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c21df91a9e4e3165686882b2eb559498224063dc633c3a63f958598108eca29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j24lb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:07:16Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-kl5s9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:45Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:45 crc kubenswrapper[4953]: I1203 00:07:45.874812 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5cmlc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7e11a1d3-ecee-4bf0-87bf-085fab97b73d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d42390b9c53efc63add8ddcebf03c976f5e5448eaa2c34e076b2dc90dbac60fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6x2sc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:07:18Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5cmlc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:45Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:45 crc kubenswrapper[4953]: I1203 00:07:45.892735 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-chmnz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a12fe6d-d25a-4b08-90da-cc9e510980a7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://711ab839cb9c0a90554dadbd45d810fc78aac09739340482fef8de8743b77283\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c982g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f802d5628a8cc32618044b9383211d5a492c56060d427f953b7793a1064029e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c982g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:07:29Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-chmnz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:45Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:45 crc kubenswrapper[4953]: I1203 00:07:45.911842 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-xshk4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bd03a758-0ca9-47d8-86f7-18b2d610df43\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47gfh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47gfh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:07:31Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-xshk4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:45Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:45 crc kubenswrapper[4953]: I1203 00:07:45.939472 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:45 crc kubenswrapper[4953]: I1203 00:07:45.939516 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:45 crc kubenswrapper[4953]: I1203 00:07:45.939526 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:45 crc kubenswrapper[4953]: I1203 00:07:45.939545 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:45 crc kubenswrapper[4953]: I1203 00:07:45.939558 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:45Z","lastTransitionTime":"2025-12-03T00:07:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:45 crc kubenswrapper[4953]: I1203 00:07:45.954144 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cf821840-5353-4620-8f73-99b56b371ab0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cae5e2994fcc24dc9b2ea73e79572615b3311edc42642b7f278fb84123989b23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://29e4b9d1b70a93fb133e0759ad148ae133766a66efeae3ff06a21dfae0e805ff\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dec43536772aa7c3f5a662fa7f3bd73b1cae2791ad454f85792287828d0ffeed\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://034e0bac6892d361700b8f33d6427b8a55e2bd3f89b14ea77a374c6be94d25e2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://761973393a93c59c23e69f41b8ac0d16abc35abd61ecbc6d51a42d94cd664233\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T00:07:09Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 00:07:03.713374 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 00:07:03.715052 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1528563048/tls.crt::/tmp/serving-cert-1528563048/tls.key\\\\\\\"\\\\nI1203 00:07:09.300052 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 00:07:09.304192 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 00:07:09.304255 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 00:07:09.304309 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 00:07:09.304323 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 00:07:09.313563 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 00:07:09.313597 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 00:07:09.313603 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 00:07:09.313609 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 00:07:09.313613 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 00:07:09.313618 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 00:07:09.313622 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 00:07:09.313709 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 00:07:09.327347 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T00:06:53Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c143d4232fb8bf7ef5999a664e431a00ab0db030f4d31a28d0bb21e00e38581\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:53Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72af6c27296eae2e275d455795c68f0d272869d63642b8353b17d466d0c901cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://72af6c27296eae2e275d455795c68f0d272869d63642b8353b17d466d0c901cf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:06:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:06:51Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:45Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:45 crc kubenswrapper[4953]: I1203 00:07:45.978765 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:10Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:45Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:45 crc kubenswrapper[4953]: I1203 00:07:45.992626 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce8b414e1fc22eb30decb6a6411b672e7f443055cb89ebb029fbf6ffc1d5dca7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:45Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:46 crc kubenswrapper[4953]: I1203 00:07:46.006183 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-6dfbn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53ffb1e9-9a2e-4869-80d7-776f4aa7ef8b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ab4205c0d38ad6a0c32077eadb1fd5fd3a279d9c8b8a2ffbe693ec2c4018370c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ct26c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7df84405eda520ad5a22d35750edf07ef3c3977a756ad5739c54f91cba9da7b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ct26c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:07:16Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-6dfbn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:46Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:46 crc kubenswrapper[4953]: I1203 00:07:46.032302 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-txffx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8816ebee-c735-4b8d-b1ec-65349707b347\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:16Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:16Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4b85224419a4075acb340e4b11095f2adb5c316e285afa4e19295c2181c0d24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0f6f645a58993f052a6f365bb808f62d99f7fea74886c5cc022b2d86112321a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a61b951a9cbab2c48c0f2df8351f97d7ff71bf3674f641fe5eac30155232b224\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd5f428de229749485267e81d34e31b708728b52474032e88460e54747f21e11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://00a2ea4a9bb870fce30e7fa845f787f184e7a19e42c95511655ea8c5f84a26d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc0a5d4cc9d614d8492626a1b114bdc52f510eb5965d842c02c083f08d4849a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dcb805f9ce285d3841e93d588fe7f3d42232382366d42b75bcfb129aefa91c1f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dcb805f9ce285d3841e93d588fe7f3d42232382366d42b75bcfb129aefa91c1f\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T00:07:44Z\\\",\\\"message\\\":\\\"mers/externalversions/factory.go:140\\\\nI1203 00:07:44.043157 6594 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1203 00:07:44.043232 6594 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 00:07:44.043450 6594 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 00:07:44.043811 6594 reflector.go:311] Stopping reflector *v1.NetworkAttachmentDefinition (0s) from github.com/k8snetworkplumbingwg/network-attachment-definition-client/pkg/client/informers/externalversions/factory.go:117\\\\nI1203 00:07:44.044028 6594 reflector.go:311] Stopping reflector *v1.AdminPolicyBasedExternalRoute (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140\\\\nI1203 00:07:44.044197 6594 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI1203 00:07:44.044257 6594 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T00:07:43Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-txffx_openshift-ovn-kubernetes(8816ebee-c735-4b8d-b1ec-65349707b347)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7deadc8e135ff1bc7c5a2f093e489bb5b4b6f166bc874c6acc2a841c5397f11e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af4c7764fca87144d01f3738ea010308744ccf50c54b7c471ccc213112de00d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af4c7764fca87144d01f3738ea010308744ccf50c54b7c471ccc213112de00d5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:07:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:07:16Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-txffx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:46Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:46 crc kubenswrapper[4953]: I1203 00:07:46.042708 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:46 crc kubenswrapper[4953]: I1203 00:07:46.042739 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:46 crc kubenswrapper[4953]: I1203 00:07:46.042748 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:46 crc kubenswrapper[4953]: I1203 00:07:46.042765 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:46 crc kubenswrapper[4953]: I1203 00:07:46.042777 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:46Z","lastTransitionTime":"2025-12-03T00:07:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:46 crc kubenswrapper[4953]: I1203 00:07:46.050258 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:46Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:46 crc kubenswrapper[4953]: I1203 00:07:46.065336 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2ed9c7fa96c8951fbc1728fe7a91fefdd999077285621df9c1abedf9ae0c449\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:46Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:46 crc kubenswrapper[4953]: I1203 00:07:46.078506 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:10Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:46Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:46 crc kubenswrapper[4953]: I1203 00:07:46.094493 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-675qk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"43a4bad8-71f4-441a-8302-aeb526a986af\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5b5794fef6477aab5deed8f642e8cc50b37eb1fd0cc6f19c003f34ad31c1e3c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gdrw4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:07:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-675qk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:46Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:46 crc kubenswrapper[4953]: I1203 00:07:46.153436 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:46 crc kubenswrapper[4953]: I1203 00:07:46.153490 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:46 crc kubenswrapper[4953]: I1203 00:07:46.153506 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:46 crc kubenswrapper[4953]: I1203 00:07:46.153530 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:46 crc kubenswrapper[4953]: I1203 00:07:46.153547 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:46Z","lastTransitionTime":"2025-12-03T00:07:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:46 crc kubenswrapper[4953]: I1203 00:07:46.259876 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:46 crc kubenswrapper[4953]: I1203 00:07:46.259934 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:46 crc kubenswrapper[4953]: I1203 00:07:46.259953 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:46 crc kubenswrapper[4953]: I1203 00:07:46.260033 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:46 crc kubenswrapper[4953]: I1203 00:07:46.260109 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:46Z","lastTransitionTime":"2025-12-03T00:07:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:46 crc kubenswrapper[4953]: I1203 00:07:46.348122 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 00:07:46 crc kubenswrapper[4953]: I1203 00:07:46.348180 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 00:07:46 crc kubenswrapper[4953]: I1203 00:07:46.348170 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 00:07:46 crc kubenswrapper[4953]: E1203 00:07:46.348345 4953 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 00:07:46 crc kubenswrapper[4953]: E1203 00:07:46.348432 4953 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 00:07:46 crc kubenswrapper[4953]: E1203 00:07:46.348549 4953 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 00:07:46 crc kubenswrapper[4953]: I1203 00:07:46.362722 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:46 crc kubenswrapper[4953]: I1203 00:07:46.362788 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:46 crc kubenswrapper[4953]: I1203 00:07:46.362812 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:46 crc kubenswrapper[4953]: I1203 00:07:46.362855 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:46 crc kubenswrapper[4953]: I1203 00:07:46.362879 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:46Z","lastTransitionTime":"2025-12-03T00:07:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:46 crc kubenswrapper[4953]: I1203 00:07:46.466427 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:46 crc kubenswrapper[4953]: I1203 00:07:46.466488 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:46 crc kubenswrapper[4953]: I1203 00:07:46.466508 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:46 crc kubenswrapper[4953]: I1203 00:07:46.466532 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:46 crc kubenswrapper[4953]: I1203 00:07:46.466549 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:46Z","lastTransitionTime":"2025-12-03T00:07:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:46 crc kubenswrapper[4953]: I1203 00:07:46.569608 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:46 crc kubenswrapper[4953]: I1203 00:07:46.570032 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:46 crc kubenswrapper[4953]: I1203 00:07:46.570060 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:46 crc kubenswrapper[4953]: I1203 00:07:46.570085 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:46 crc kubenswrapper[4953]: I1203 00:07:46.570102 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:46Z","lastTransitionTime":"2025-12-03T00:07:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:46 crc kubenswrapper[4953]: I1203 00:07:46.672501 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:46 crc kubenswrapper[4953]: I1203 00:07:46.672565 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:46 crc kubenswrapper[4953]: I1203 00:07:46.672584 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:46 crc kubenswrapper[4953]: I1203 00:07:46.672611 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:46 crc kubenswrapper[4953]: I1203 00:07:46.672631 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:46Z","lastTransitionTime":"2025-12-03T00:07:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:46 crc kubenswrapper[4953]: I1203 00:07:46.776254 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:46 crc kubenswrapper[4953]: I1203 00:07:46.776307 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:46 crc kubenswrapper[4953]: I1203 00:07:46.776324 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:46 crc kubenswrapper[4953]: I1203 00:07:46.776352 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:46 crc kubenswrapper[4953]: I1203 00:07:46.776368 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:46Z","lastTransitionTime":"2025-12-03T00:07:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:46 crc kubenswrapper[4953]: I1203 00:07:46.879105 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:46 crc kubenswrapper[4953]: I1203 00:07:46.879174 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:46 crc kubenswrapper[4953]: I1203 00:07:46.879194 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:46 crc kubenswrapper[4953]: I1203 00:07:46.879217 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:46 crc kubenswrapper[4953]: I1203 00:07:46.879234 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:46Z","lastTransitionTime":"2025-12-03T00:07:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:46 crc kubenswrapper[4953]: I1203 00:07:46.958824 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/bd03a758-0ca9-47d8-86f7-18b2d610df43-metrics-certs\") pod \"network-metrics-daemon-xshk4\" (UID: \"bd03a758-0ca9-47d8-86f7-18b2d610df43\") " pod="openshift-multus/network-metrics-daemon-xshk4" Dec 03 00:07:46 crc kubenswrapper[4953]: E1203 00:07:46.959033 4953 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 03 00:07:46 crc kubenswrapper[4953]: E1203 00:07:46.959147 4953 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/bd03a758-0ca9-47d8-86f7-18b2d610df43-metrics-certs podName:bd03a758-0ca9-47d8-86f7-18b2d610df43 nodeName:}" failed. No retries permitted until 2025-12-03 00:08:02.959126632 +0000 UTC m=+71.943407425 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/bd03a758-0ca9-47d8-86f7-18b2d610df43-metrics-certs") pod "network-metrics-daemon-xshk4" (UID: "bd03a758-0ca9-47d8-86f7-18b2d610df43") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 03 00:07:46 crc kubenswrapper[4953]: I1203 00:07:46.982928 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:46 crc kubenswrapper[4953]: I1203 00:07:46.983044 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:46 crc kubenswrapper[4953]: I1203 00:07:46.983073 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:46 crc kubenswrapper[4953]: I1203 00:07:46.983104 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:46 crc kubenswrapper[4953]: I1203 00:07:46.983126 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:46Z","lastTransitionTime":"2025-12-03T00:07:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:47 crc kubenswrapper[4953]: I1203 00:07:47.087233 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:47 crc kubenswrapper[4953]: I1203 00:07:47.088105 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:47 crc kubenswrapper[4953]: I1203 00:07:47.088123 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:47 crc kubenswrapper[4953]: I1203 00:07:47.088145 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:47 crc kubenswrapper[4953]: I1203 00:07:47.088157 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:47Z","lastTransitionTime":"2025-12-03T00:07:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:47 crc kubenswrapper[4953]: I1203 00:07:47.191865 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:47 crc kubenswrapper[4953]: I1203 00:07:47.191948 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:47 crc kubenswrapper[4953]: I1203 00:07:47.192060 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:47 crc kubenswrapper[4953]: I1203 00:07:47.192097 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:47 crc kubenswrapper[4953]: I1203 00:07:47.192121 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:47Z","lastTransitionTime":"2025-12-03T00:07:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:47 crc kubenswrapper[4953]: I1203 00:07:47.295580 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:47 crc kubenswrapper[4953]: I1203 00:07:47.295656 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:47 crc kubenswrapper[4953]: I1203 00:07:47.295678 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:47 crc kubenswrapper[4953]: I1203 00:07:47.295703 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:47 crc kubenswrapper[4953]: I1203 00:07:47.295722 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:47Z","lastTransitionTime":"2025-12-03T00:07:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:47 crc kubenswrapper[4953]: I1203 00:07:47.348303 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xshk4" Dec 03 00:07:47 crc kubenswrapper[4953]: E1203 00:07:47.348509 4953 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xshk4" podUID="bd03a758-0ca9-47d8-86f7-18b2d610df43" Dec 03 00:07:47 crc kubenswrapper[4953]: I1203 00:07:47.399328 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:47 crc kubenswrapper[4953]: I1203 00:07:47.399397 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:47 crc kubenswrapper[4953]: I1203 00:07:47.399419 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:47 crc kubenswrapper[4953]: I1203 00:07:47.399449 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:47 crc kubenswrapper[4953]: I1203 00:07:47.399471 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:47Z","lastTransitionTime":"2025-12-03T00:07:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:47 crc kubenswrapper[4953]: I1203 00:07:47.503399 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:47 crc kubenswrapper[4953]: I1203 00:07:47.503458 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:47 crc kubenswrapper[4953]: I1203 00:07:47.503475 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:47 crc kubenswrapper[4953]: I1203 00:07:47.503499 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:47 crc kubenswrapper[4953]: I1203 00:07:47.503517 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:47Z","lastTransitionTime":"2025-12-03T00:07:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:47 crc kubenswrapper[4953]: I1203 00:07:47.607088 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:47 crc kubenswrapper[4953]: I1203 00:07:47.607176 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:47 crc kubenswrapper[4953]: I1203 00:07:47.607213 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:47 crc kubenswrapper[4953]: I1203 00:07:47.607245 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:47 crc kubenswrapper[4953]: I1203 00:07:47.607269 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:47Z","lastTransitionTime":"2025-12-03T00:07:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:47 crc kubenswrapper[4953]: I1203 00:07:47.710183 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:47 crc kubenswrapper[4953]: I1203 00:07:47.710251 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:47 crc kubenswrapper[4953]: I1203 00:07:47.710269 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:47 crc kubenswrapper[4953]: I1203 00:07:47.710295 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:47 crc kubenswrapper[4953]: I1203 00:07:47.710312 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:47Z","lastTransitionTime":"2025-12-03T00:07:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:47 crc kubenswrapper[4953]: I1203 00:07:47.813404 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:47 crc kubenswrapper[4953]: I1203 00:07:47.813489 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:47 crc kubenswrapper[4953]: I1203 00:07:47.813516 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:47 crc kubenswrapper[4953]: I1203 00:07:47.813550 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:47 crc kubenswrapper[4953]: I1203 00:07:47.813573 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:47Z","lastTransitionTime":"2025-12-03T00:07:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:47 crc kubenswrapper[4953]: I1203 00:07:47.916193 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:47 crc kubenswrapper[4953]: I1203 00:07:47.916272 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:47 crc kubenswrapper[4953]: I1203 00:07:47.916289 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:47 crc kubenswrapper[4953]: I1203 00:07:47.916304 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:47 crc kubenswrapper[4953]: I1203 00:07:47.916316 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:47Z","lastTransitionTime":"2025-12-03T00:07:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:48 crc kubenswrapper[4953]: I1203 00:07:48.019879 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:48 crc kubenswrapper[4953]: I1203 00:07:48.019941 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:48 crc kubenswrapper[4953]: I1203 00:07:48.019960 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:48 crc kubenswrapper[4953]: I1203 00:07:48.020033 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:48 crc kubenswrapper[4953]: I1203 00:07:48.020057 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:48Z","lastTransitionTime":"2025-12-03T00:07:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:48 crc kubenswrapper[4953]: I1203 00:07:48.123717 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:48 crc kubenswrapper[4953]: I1203 00:07:48.123825 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:48 crc kubenswrapper[4953]: I1203 00:07:48.123841 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:48 crc kubenswrapper[4953]: I1203 00:07:48.123865 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:48 crc kubenswrapper[4953]: I1203 00:07:48.123884 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:48Z","lastTransitionTime":"2025-12-03T00:07:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:48 crc kubenswrapper[4953]: I1203 00:07:48.226549 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:48 crc kubenswrapper[4953]: I1203 00:07:48.226599 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:48 crc kubenswrapper[4953]: I1203 00:07:48.226611 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:48 crc kubenswrapper[4953]: I1203 00:07:48.226627 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:48 crc kubenswrapper[4953]: I1203 00:07:48.226641 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:48Z","lastTransitionTime":"2025-12-03T00:07:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:48 crc kubenswrapper[4953]: I1203 00:07:48.329122 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:48 crc kubenswrapper[4953]: I1203 00:07:48.329676 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:48 crc kubenswrapper[4953]: I1203 00:07:48.329691 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:48 crc kubenswrapper[4953]: I1203 00:07:48.329714 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:48 crc kubenswrapper[4953]: I1203 00:07:48.329727 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:48Z","lastTransitionTime":"2025-12-03T00:07:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:48 crc kubenswrapper[4953]: I1203 00:07:48.347730 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 00:07:48 crc kubenswrapper[4953]: I1203 00:07:48.347730 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 00:07:48 crc kubenswrapper[4953]: E1203 00:07:48.347919 4953 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 00:07:48 crc kubenswrapper[4953]: I1203 00:07:48.347759 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 00:07:48 crc kubenswrapper[4953]: E1203 00:07:48.348198 4953 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 00:07:48 crc kubenswrapper[4953]: E1203 00:07:48.348324 4953 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 00:07:48 crc kubenswrapper[4953]: I1203 00:07:48.433667 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:48 crc kubenswrapper[4953]: I1203 00:07:48.433752 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:48 crc kubenswrapper[4953]: I1203 00:07:48.433764 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:48 crc kubenswrapper[4953]: I1203 00:07:48.433785 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:48 crc kubenswrapper[4953]: I1203 00:07:48.433818 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:48Z","lastTransitionTime":"2025-12-03T00:07:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:48 crc kubenswrapper[4953]: I1203 00:07:48.538493 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:48 crc kubenswrapper[4953]: I1203 00:07:48.538569 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:48 crc kubenswrapper[4953]: I1203 00:07:48.538584 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:48 crc kubenswrapper[4953]: I1203 00:07:48.538614 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:48 crc kubenswrapper[4953]: I1203 00:07:48.538631 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:48Z","lastTransitionTime":"2025-12-03T00:07:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:48 crc kubenswrapper[4953]: I1203 00:07:48.642404 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:48 crc kubenswrapper[4953]: I1203 00:07:48.642472 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:48 crc kubenswrapper[4953]: I1203 00:07:48.642490 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:48 crc kubenswrapper[4953]: I1203 00:07:48.642516 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:48 crc kubenswrapper[4953]: I1203 00:07:48.642761 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:48Z","lastTransitionTime":"2025-12-03T00:07:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:48 crc kubenswrapper[4953]: I1203 00:07:48.746259 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:48 crc kubenswrapper[4953]: I1203 00:07:48.746304 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:48 crc kubenswrapper[4953]: I1203 00:07:48.746313 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:48 crc kubenswrapper[4953]: I1203 00:07:48.746334 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:48 crc kubenswrapper[4953]: I1203 00:07:48.746346 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:48Z","lastTransitionTime":"2025-12-03T00:07:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:48 crc kubenswrapper[4953]: I1203 00:07:48.849550 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:48 crc kubenswrapper[4953]: I1203 00:07:48.849596 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:48 crc kubenswrapper[4953]: I1203 00:07:48.849610 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:48 crc kubenswrapper[4953]: I1203 00:07:48.849628 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:48 crc kubenswrapper[4953]: I1203 00:07:48.849641 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:48Z","lastTransitionTime":"2025-12-03T00:07:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:48 crc kubenswrapper[4953]: I1203 00:07:48.952940 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:48 crc kubenswrapper[4953]: I1203 00:07:48.953010 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:48 crc kubenswrapper[4953]: I1203 00:07:48.953024 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:48 crc kubenswrapper[4953]: I1203 00:07:48.953045 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:48 crc kubenswrapper[4953]: I1203 00:07:48.953058 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:48Z","lastTransitionTime":"2025-12-03T00:07:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:49 crc kubenswrapper[4953]: I1203 00:07:49.056163 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:49 crc kubenswrapper[4953]: I1203 00:07:49.056262 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:49 crc kubenswrapper[4953]: I1203 00:07:49.056282 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:49 crc kubenswrapper[4953]: I1203 00:07:49.056306 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:49 crc kubenswrapper[4953]: I1203 00:07:49.056324 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:49Z","lastTransitionTime":"2025-12-03T00:07:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:49 crc kubenswrapper[4953]: I1203 00:07:49.159765 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:49 crc kubenswrapper[4953]: I1203 00:07:49.159838 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:49 crc kubenswrapper[4953]: I1203 00:07:49.159862 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:49 crc kubenswrapper[4953]: I1203 00:07:49.159894 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:49 crc kubenswrapper[4953]: I1203 00:07:49.159917 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:49Z","lastTransitionTime":"2025-12-03T00:07:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:49 crc kubenswrapper[4953]: I1203 00:07:49.263163 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:49 crc kubenswrapper[4953]: I1203 00:07:49.263230 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:49 crc kubenswrapper[4953]: I1203 00:07:49.263248 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:49 crc kubenswrapper[4953]: I1203 00:07:49.263271 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:49 crc kubenswrapper[4953]: I1203 00:07:49.263289 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:49Z","lastTransitionTime":"2025-12-03T00:07:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:49 crc kubenswrapper[4953]: I1203 00:07:49.348031 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xshk4" Dec 03 00:07:49 crc kubenswrapper[4953]: E1203 00:07:49.348302 4953 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xshk4" podUID="bd03a758-0ca9-47d8-86f7-18b2d610df43" Dec 03 00:07:49 crc kubenswrapper[4953]: I1203 00:07:49.366173 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:49 crc kubenswrapper[4953]: I1203 00:07:49.366240 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:49 crc kubenswrapper[4953]: I1203 00:07:49.366258 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:49 crc kubenswrapper[4953]: I1203 00:07:49.366284 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:49 crc kubenswrapper[4953]: I1203 00:07:49.366303 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:49Z","lastTransitionTime":"2025-12-03T00:07:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:49 crc kubenswrapper[4953]: I1203 00:07:49.469784 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:49 crc kubenswrapper[4953]: I1203 00:07:49.469860 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:49 crc kubenswrapper[4953]: I1203 00:07:49.469879 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:49 crc kubenswrapper[4953]: I1203 00:07:49.469906 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:49 crc kubenswrapper[4953]: I1203 00:07:49.469924 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:49Z","lastTransitionTime":"2025-12-03T00:07:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:49 crc kubenswrapper[4953]: I1203 00:07:49.572797 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:49 crc kubenswrapper[4953]: I1203 00:07:49.572860 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:49 crc kubenswrapper[4953]: I1203 00:07:49.572880 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:49 crc kubenswrapper[4953]: I1203 00:07:49.572907 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:49 crc kubenswrapper[4953]: I1203 00:07:49.572926 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:49Z","lastTransitionTime":"2025-12-03T00:07:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:49 crc kubenswrapper[4953]: I1203 00:07:49.676421 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:49 crc kubenswrapper[4953]: I1203 00:07:49.676462 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:49 crc kubenswrapper[4953]: I1203 00:07:49.676476 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:49 crc kubenswrapper[4953]: I1203 00:07:49.676497 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:49 crc kubenswrapper[4953]: I1203 00:07:49.676513 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:49Z","lastTransitionTime":"2025-12-03T00:07:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:49 crc kubenswrapper[4953]: I1203 00:07:49.779267 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:49 crc kubenswrapper[4953]: I1203 00:07:49.779326 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:49 crc kubenswrapper[4953]: I1203 00:07:49.779345 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:49 crc kubenswrapper[4953]: I1203 00:07:49.779369 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:49 crc kubenswrapper[4953]: I1203 00:07:49.779386 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:49Z","lastTransitionTime":"2025-12-03T00:07:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:49 crc kubenswrapper[4953]: I1203 00:07:49.881580 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:49 crc kubenswrapper[4953]: I1203 00:07:49.881611 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:49 crc kubenswrapper[4953]: I1203 00:07:49.881619 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:49 crc kubenswrapper[4953]: I1203 00:07:49.881632 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:49 crc kubenswrapper[4953]: I1203 00:07:49.881641 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:49Z","lastTransitionTime":"2025-12-03T00:07:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:49 crc kubenswrapper[4953]: I1203 00:07:49.984659 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:49 crc kubenswrapper[4953]: I1203 00:07:49.984694 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:49 crc kubenswrapper[4953]: I1203 00:07:49.984703 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:49 crc kubenswrapper[4953]: I1203 00:07:49.984718 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:49 crc kubenswrapper[4953]: I1203 00:07:49.984727 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:49Z","lastTransitionTime":"2025-12-03T00:07:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:50 crc kubenswrapper[4953]: I1203 00:07:50.088312 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:50 crc kubenswrapper[4953]: I1203 00:07:50.088366 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:50 crc kubenswrapper[4953]: I1203 00:07:50.088383 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:50 crc kubenswrapper[4953]: I1203 00:07:50.088406 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:50 crc kubenswrapper[4953]: I1203 00:07:50.088423 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:50Z","lastTransitionTime":"2025-12-03T00:07:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:50 crc kubenswrapper[4953]: I1203 00:07:50.192244 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:50 crc kubenswrapper[4953]: I1203 00:07:50.192317 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:50 crc kubenswrapper[4953]: I1203 00:07:50.192335 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:50 crc kubenswrapper[4953]: I1203 00:07:50.192358 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:50 crc kubenswrapper[4953]: I1203 00:07:50.192381 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:50Z","lastTransitionTime":"2025-12-03T00:07:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:50 crc kubenswrapper[4953]: I1203 00:07:50.295191 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:50 crc kubenswrapper[4953]: I1203 00:07:50.295252 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:50 crc kubenswrapper[4953]: I1203 00:07:50.295275 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:50 crc kubenswrapper[4953]: I1203 00:07:50.295305 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:50 crc kubenswrapper[4953]: I1203 00:07:50.295328 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:50Z","lastTransitionTime":"2025-12-03T00:07:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:50 crc kubenswrapper[4953]: I1203 00:07:50.347202 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 00:07:50 crc kubenswrapper[4953]: I1203 00:07:50.347269 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 00:07:50 crc kubenswrapper[4953]: I1203 00:07:50.347202 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 00:07:50 crc kubenswrapper[4953]: E1203 00:07:50.347365 4953 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 00:07:50 crc kubenswrapper[4953]: E1203 00:07:50.347567 4953 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 00:07:50 crc kubenswrapper[4953]: E1203 00:07:50.347693 4953 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 00:07:50 crc kubenswrapper[4953]: I1203 00:07:50.398637 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:50 crc kubenswrapper[4953]: I1203 00:07:50.398707 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:50 crc kubenswrapper[4953]: I1203 00:07:50.398731 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:50 crc kubenswrapper[4953]: I1203 00:07:50.398762 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:50 crc kubenswrapper[4953]: I1203 00:07:50.398785 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:50Z","lastTransitionTime":"2025-12-03T00:07:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:50 crc kubenswrapper[4953]: I1203 00:07:50.501488 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:50 crc kubenswrapper[4953]: I1203 00:07:50.501572 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:50 crc kubenswrapper[4953]: I1203 00:07:50.501589 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:50 crc kubenswrapper[4953]: I1203 00:07:50.501612 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:50 crc kubenswrapper[4953]: I1203 00:07:50.501628 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:50Z","lastTransitionTime":"2025-12-03T00:07:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:50 crc kubenswrapper[4953]: I1203 00:07:50.605275 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:50 crc kubenswrapper[4953]: I1203 00:07:50.605346 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:50 crc kubenswrapper[4953]: I1203 00:07:50.605371 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:50 crc kubenswrapper[4953]: I1203 00:07:50.605400 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:50 crc kubenswrapper[4953]: I1203 00:07:50.605421 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:50Z","lastTransitionTime":"2025-12-03T00:07:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:50 crc kubenswrapper[4953]: I1203 00:07:50.708196 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:50 crc kubenswrapper[4953]: I1203 00:07:50.708233 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:50 crc kubenswrapper[4953]: I1203 00:07:50.708244 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:50 crc kubenswrapper[4953]: I1203 00:07:50.708261 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:50 crc kubenswrapper[4953]: I1203 00:07:50.708273 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:50Z","lastTransitionTime":"2025-12-03T00:07:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:50 crc kubenswrapper[4953]: I1203 00:07:50.811523 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:50 crc kubenswrapper[4953]: I1203 00:07:50.811577 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:50 crc kubenswrapper[4953]: I1203 00:07:50.811594 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:50 crc kubenswrapper[4953]: I1203 00:07:50.811618 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:50 crc kubenswrapper[4953]: I1203 00:07:50.811636 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:50Z","lastTransitionTime":"2025-12-03T00:07:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:50 crc kubenswrapper[4953]: I1203 00:07:50.914769 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:50 crc kubenswrapper[4953]: I1203 00:07:50.914817 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:50 crc kubenswrapper[4953]: I1203 00:07:50.914834 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:50 crc kubenswrapper[4953]: I1203 00:07:50.914858 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:50 crc kubenswrapper[4953]: I1203 00:07:50.914876 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:50Z","lastTransitionTime":"2025-12-03T00:07:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:51 crc kubenswrapper[4953]: I1203 00:07:51.017833 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:51 crc kubenswrapper[4953]: I1203 00:07:51.017903 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:51 crc kubenswrapper[4953]: I1203 00:07:51.017929 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:51 crc kubenswrapper[4953]: I1203 00:07:51.018013 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:51 crc kubenswrapper[4953]: I1203 00:07:51.018043 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:51Z","lastTransitionTime":"2025-12-03T00:07:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:51 crc kubenswrapper[4953]: I1203 00:07:51.121221 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:51 crc kubenswrapper[4953]: I1203 00:07:51.121287 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:51 crc kubenswrapper[4953]: I1203 00:07:51.121306 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:51 crc kubenswrapper[4953]: I1203 00:07:51.121329 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:51 crc kubenswrapper[4953]: I1203 00:07:51.121346 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:51Z","lastTransitionTime":"2025-12-03T00:07:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:51 crc kubenswrapper[4953]: I1203 00:07:51.224932 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:51 crc kubenswrapper[4953]: I1203 00:07:51.225390 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:51 crc kubenswrapper[4953]: I1203 00:07:51.225597 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:51 crc kubenswrapper[4953]: I1203 00:07:51.225766 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:51 crc kubenswrapper[4953]: I1203 00:07:51.225949 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:51Z","lastTransitionTime":"2025-12-03T00:07:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:51 crc kubenswrapper[4953]: I1203 00:07:51.329366 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:51 crc kubenswrapper[4953]: I1203 00:07:51.329427 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:51 crc kubenswrapper[4953]: I1203 00:07:51.329446 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:51 crc kubenswrapper[4953]: I1203 00:07:51.329471 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:51 crc kubenswrapper[4953]: I1203 00:07:51.329491 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:51Z","lastTransitionTime":"2025-12-03T00:07:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:51 crc kubenswrapper[4953]: I1203 00:07:51.347218 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xshk4" Dec 03 00:07:51 crc kubenswrapper[4953]: E1203 00:07:51.347364 4953 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xshk4" podUID="bd03a758-0ca9-47d8-86f7-18b2d610df43" Dec 03 00:07:51 crc kubenswrapper[4953]: I1203 00:07:51.371047 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:10Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:51Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:51 crc kubenswrapper[4953]: I1203 00:07:51.391170 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-675qk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"43a4bad8-71f4-441a-8302-aeb526a986af\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5b5794fef6477aab5deed8f642e8cc50b37eb1fd0cc6f19c003f34ad31c1e3c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gdrw4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:07:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-675qk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:51Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:51 crc kubenswrapper[4953]: I1203 00:07:51.410735 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f94f8c51-c665-4ca3-bf17-7a04ea5c7ade\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72c5cd7dc1c3a8adfedab6ba43a59d80b3efe65f2bb56be95e76cc86a409dcf3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2810ffc1b141fa7da811406053db9d72c708edf5f2bb1dd75e4d52ab35fd24c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c67b4277b20a9f1287f50ef82d0785c9d02543a8e7ed26e0413eae332b866db7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://258d78e1ad9d3dcda6976909ae3fe71aecead5b995fc66b3fd6e2722d80b5ab8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:06:51Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:51Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:51 crc kubenswrapper[4953]: I1203 00:07:51.429459 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5652d0f3-3fd7-4af6-8941-9119e1177641\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2030aef5b7170991268fa5dc7e7ed8ce1fbe3a2765165a10c25251426575330f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ab6052f1abf17225155ce1bd86afa742e6e78f507b7deedd7161cd8fba9120b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://650b6c79cda2e1768ba1d8763db799e210ce2748893512068d25e60f821937ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d1d60162f7f6885d4f51b4ab44163cfae6d2d04355b1dd710e26b022450005df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d1d60162f7f6885d4f51b4ab44163cfae6d2d04355b1dd710e26b022450005df\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:06:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:06:51Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:51Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:51 crc kubenswrapper[4953]: I1203 00:07:51.432524 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:51 crc kubenswrapper[4953]: I1203 00:07:51.434659 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:51 crc kubenswrapper[4953]: I1203 00:07:51.434864 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:51 crc kubenswrapper[4953]: I1203 00:07:51.435068 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:51 crc kubenswrapper[4953]: I1203 00:07:51.435251 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:51Z","lastTransitionTime":"2025-12-03T00:07:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:51 crc kubenswrapper[4953]: I1203 00:07:51.448595 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09657a9f448a4c92a759c73a0a892076d7eb46393e6fd565e5f73f34c71aac23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5af5eac4a58870f63497f99b503cfa5fb922278095e0332ec50376925ecf2109\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:51Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:51 crc kubenswrapper[4953]: I1203 00:07:51.477714 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-wj6rz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"96aa67a0-15dd-420d-a1d5-ed239d7df1fb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01815cfd3f58c3f3f5f2a894c1790e0b69c5060c48dc0b1deff050b6ad7af916\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kf62m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b12bfb5bdf9fd4f2c50a335fc82df6ee0440eaf8741897da03aa2dae368cc6b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b12bfb5bdf9fd4f2c50a335fc82df6ee0440eaf8741897da03aa2dae368cc6b4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:07:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kf62m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83aaa5fa28110bc0ef05aff75cd1442bc9bbdd33b822ae2ef22d4e4ab83238bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://83aaa5fa28110bc0ef05aff75cd1442bc9bbdd33b822ae2ef22d4e4ab83238bc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:07:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:07:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kf62m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ab6213a33a19fedebce6d74a0ebbb3b3a2de5687504e21cc24093cc42ade635\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ab6213a33a19fedebce6d74a0ebbb3b3a2de5687504e21cc24093cc42ade635\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:07:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:07:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kf62m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://de480488843afa6b8dede1e1b42b5181dca1993e2f39dc1b461a18a57e3668e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://de480488843afa6b8dede1e1b42b5181dca1993e2f39dc1b461a18a57e3668e1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:07:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kf62m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://97be4ba692bc61078d4b0ba39fb0c12772548c6c9251ff3356ea702f151223cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://97be4ba692bc61078d4b0ba39fb0c12772548c6c9251ff3356ea702f151223cd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:07:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kf62m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1a1f009fcbd8af9217db1dc97f662719f8c1db232df0e2804e44c881abb1df2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b1a1f009fcbd8af9217db1dc97f662719f8c1db232df0e2804e44c881abb1df2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:07:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:07:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kf62m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:07:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-wj6rz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:51Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:51 crc kubenswrapper[4953]: I1203 00:07:51.489996 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-kl5s9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d85e5c97-16ad-4bd7-9e73-b7f12678be97\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c21df91a9e4e3165686882b2eb559498224063dc633c3a63f958598108eca29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j24lb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:07:16Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-kl5s9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:51Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:51 crc kubenswrapper[4953]: I1203 00:07:51.502039 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5cmlc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7e11a1d3-ecee-4bf0-87bf-085fab97b73d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d42390b9c53efc63add8ddcebf03c976f5e5448eaa2c34e076b2dc90dbac60fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6x2sc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:07:18Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5cmlc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:51Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:51 crc kubenswrapper[4953]: I1203 00:07:51.516962 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-chmnz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a12fe6d-d25a-4b08-90da-cc9e510980a7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://711ab839cb9c0a90554dadbd45d810fc78aac09739340482fef8de8743b77283\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c982g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f802d5628a8cc32618044b9383211d5a492c56060d427f953b7793a1064029e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c982g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:07:29Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-chmnz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:51Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:51 crc kubenswrapper[4953]: I1203 00:07:51.530951 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-xshk4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bd03a758-0ca9-47d8-86f7-18b2d610df43\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47gfh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47gfh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:07:31Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-xshk4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:51Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:51 crc kubenswrapper[4953]: I1203 00:07:51.537688 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:51 crc kubenswrapper[4953]: I1203 00:07:51.537735 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:51 crc kubenswrapper[4953]: I1203 00:07:51.537751 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:51 crc kubenswrapper[4953]: I1203 00:07:51.537770 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:51 crc kubenswrapper[4953]: I1203 00:07:51.537785 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:51Z","lastTransitionTime":"2025-12-03T00:07:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:51 crc kubenswrapper[4953]: I1203 00:07:51.556188 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cf821840-5353-4620-8f73-99b56b371ab0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cae5e2994fcc24dc9b2ea73e79572615b3311edc42642b7f278fb84123989b23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://29e4b9d1b70a93fb133e0759ad148ae133766a66efeae3ff06a21dfae0e805ff\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dec43536772aa7c3f5a662fa7f3bd73b1cae2791ad454f85792287828d0ffeed\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://034e0bac6892d361700b8f33d6427b8a55e2bd3f89b14ea77a374c6be94d25e2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://761973393a93c59c23e69f41b8ac0d16abc35abd61ecbc6d51a42d94cd664233\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T00:07:09Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 00:07:03.713374 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 00:07:03.715052 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1528563048/tls.crt::/tmp/serving-cert-1528563048/tls.key\\\\\\\"\\\\nI1203 00:07:09.300052 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 00:07:09.304192 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 00:07:09.304255 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 00:07:09.304309 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 00:07:09.304323 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 00:07:09.313563 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 00:07:09.313597 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 00:07:09.313603 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 00:07:09.313609 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 00:07:09.313613 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 00:07:09.313618 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 00:07:09.313622 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 00:07:09.313709 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 00:07:09.327347 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T00:06:53Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c143d4232fb8bf7ef5999a664e431a00ab0db030f4d31a28d0bb21e00e38581\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:53Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72af6c27296eae2e275d455795c68f0d272869d63642b8353b17d466d0c901cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://72af6c27296eae2e275d455795c68f0d272869d63642b8353b17d466d0c901cf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:06:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:06:51Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:51Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:51 crc kubenswrapper[4953]: I1203 00:07:51.577004 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:10Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:51Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:51 crc kubenswrapper[4953]: I1203 00:07:51.597608 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce8b414e1fc22eb30decb6a6411b672e7f443055cb89ebb029fbf6ffc1d5dca7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:51Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:51 crc kubenswrapper[4953]: I1203 00:07:51.615952 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-6dfbn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53ffb1e9-9a2e-4869-80d7-776f4aa7ef8b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ab4205c0d38ad6a0c32077eadb1fd5fd3a279d9c8b8a2ffbe693ec2c4018370c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ct26c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7df84405eda520ad5a22d35750edf07ef3c3977a756ad5739c54f91cba9da7b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ct26c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:07:16Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-6dfbn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:51Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:51 crc kubenswrapper[4953]: I1203 00:07:51.644581 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:51 crc kubenswrapper[4953]: I1203 00:07:51.644640 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:51 crc kubenswrapper[4953]: I1203 00:07:51.644657 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:51 crc kubenswrapper[4953]: I1203 00:07:51.644678 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:51 crc kubenswrapper[4953]: I1203 00:07:51.644695 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:51Z","lastTransitionTime":"2025-12-03T00:07:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:51 crc kubenswrapper[4953]: I1203 00:07:51.646136 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-txffx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8816ebee-c735-4b8d-b1ec-65349707b347\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:16Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:16Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4b85224419a4075acb340e4b11095f2adb5c316e285afa4e19295c2181c0d24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0f6f645a58993f052a6f365bb808f62d99f7fea74886c5cc022b2d86112321a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a61b951a9cbab2c48c0f2df8351f97d7ff71bf3674f641fe5eac30155232b224\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd5f428de229749485267e81d34e31b708728b52474032e88460e54747f21e11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://00a2ea4a9bb870fce30e7fa845f787f184e7a19e42c95511655ea8c5f84a26d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc0a5d4cc9d614d8492626a1b114bdc52f510eb5965d842c02c083f08d4849a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dcb805f9ce285d3841e93d588fe7f3d42232382366d42b75bcfb129aefa91c1f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dcb805f9ce285d3841e93d588fe7f3d42232382366d42b75bcfb129aefa91c1f\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T00:07:44Z\\\",\\\"message\\\":\\\"mers/externalversions/factory.go:140\\\\nI1203 00:07:44.043157 6594 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1203 00:07:44.043232 6594 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 00:07:44.043450 6594 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 00:07:44.043811 6594 reflector.go:311] Stopping reflector *v1.NetworkAttachmentDefinition (0s) from github.com/k8snetworkplumbingwg/network-attachment-definition-client/pkg/client/informers/externalversions/factory.go:117\\\\nI1203 00:07:44.044028 6594 reflector.go:311] Stopping reflector *v1.AdminPolicyBasedExternalRoute (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140\\\\nI1203 00:07:44.044197 6594 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI1203 00:07:44.044257 6594 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T00:07:43Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-txffx_openshift-ovn-kubernetes(8816ebee-c735-4b8d-b1ec-65349707b347)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7deadc8e135ff1bc7c5a2f093e489bb5b4b6f166bc874c6acc2a841c5397f11e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af4c7764fca87144d01f3738ea010308744ccf50c54b7c471ccc213112de00d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af4c7764fca87144d01f3738ea010308744ccf50c54b7c471ccc213112de00d5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:07:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:07:16Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-txffx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:51Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:51 crc kubenswrapper[4953]: I1203 00:07:51.660701 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:51Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:51 crc kubenswrapper[4953]: I1203 00:07:51.675045 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2ed9c7fa96c8951fbc1728fe7a91fefdd999077285621df9c1abedf9ae0c449\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:51Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:51 crc kubenswrapper[4953]: I1203 00:07:51.748212 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:51 crc kubenswrapper[4953]: I1203 00:07:51.748266 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:51 crc kubenswrapper[4953]: I1203 00:07:51.748280 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:51 crc kubenswrapper[4953]: I1203 00:07:51.748300 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:51 crc kubenswrapper[4953]: I1203 00:07:51.748315 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:51Z","lastTransitionTime":"2025-12-03T00:07:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:51 crc kubenswrapper[4953]: I1203 00:07:51.851590 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:51 crc kubenswrapper[4953]: I1203 00:07:51.851655 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:51 crc kubenswrapper[4953]: I1203 00:07:51.851675 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:51 crc kubenswrapper[4953]: I1203 00:07:51.851702 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:51 crc kubenswrapper[4953]: I1203 00:07:51.851722 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:51Z","lastTransitionTime":"2025-12-03T00:07:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:51 crc kubenswrapper[4953]: I1203 00:07:51.955217 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:51 crc kubenswrapper[4953]: I1203 00:07:51.955279 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:51 crc kubenswrapper[4953]: I1203 00:07:51.955296 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:51 crc kubenswrapper[4953]: I1203 00:07:51.955319 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:51 crc kubenswrapper[4953]: I1203 00:07:51.955336 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:51Z","lastTransitionTime":"2025-12-03T00:07:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:52 crc kubenswrapper[4953]: I1203 00:07:52.058643 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:52 crc kubenswrapper[4953]: I1203 00:07:52.058729 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:52 crc kubenswrapper[4953]: I1203 00:07:52.058752 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:52 crc kubenswrapper[4953]: I1203 00:07:52.058787 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:52 crc kubenswrapper[4953]: I1203 00:07:52.058808 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:52Z","lastTransitionTime":"2025-12-03T00:07:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:52 crc kubenswrapper[4953]: I1203 00:07:52.162149 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:52 crc kubenswrapper[4953]: I1203 00:07:52.162247 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:52 crc kubenswrapper[4953]: I1203 00:07:52.162266 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:52 crc kubenswrapper[4953]: I1203 00:07:52.162289 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:52 crc kubenswrapper[4953]: I1203 00:07:52.162306 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:52Z","lastTransitionTime":"2025-12-03T00:07:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:52 crc kubenswrapper[4953]: I1203 00:07:52.264924 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:52 crc kubenswrapper[4953]: I1203 00:07:52.265023 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:52 crc kubenswrapper[4953]: I1203 00:07:52.265050 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:52 crc kubenswrapper[4953]: I1203 00:07:52.265077 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:52 crc kubenswrapper[4953]: I1203 00:07:52.265094 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:52Z","lastTransitionTime":"2025-12-03T00:07:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:52 crc kubenswrapper[4953]: I1203 00:07:52.347642 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 00:07:52 crc kubenswrapper[4953]: I1203 00:07:52.347725 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 00:07:52 crc kubenswrapper[4953]: I1203 00:07:52.347653 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 00:07:52 crc kubenswrapper[4953]: E1203 00:07:52.347820 4953 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 00:07:52 crc kubenswrapper[4953]: E1203 00:07:52.347900 4953 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 00:07:52 crc kubenswrapper[4953]: E1203 00:07:52.348041 4953 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 00:07:52 crc kubenswrapper[4953]: I1203 00:07:52.367710 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:52 crc kubenswrapper[4953]: I1203 00:07:52.367752 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:52 crc kubenswrapper[4953]: I1203 00:07:52.367767 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:52 crc kubenswrapper[4953]: I1203 00:07:52.367791 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:52 crc kubenswrapper[4953]: I1203 00:07:52.367806 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:52Z","lastTransitionTime":"2025-12-03T00:07:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:52 crc kubenswrapper[4953]: I1203 00:07:52.470510 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:52 crc kubenswrapper[4953]: I1203 00:07:52.470602 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:52 crc kubenswrapper[4953]: I1203 00:07:52.470631 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:52 crc kubenswrapper[4953]: I1203 00:07:52.470667 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:52 crc kubenswrapper[4953]: I1203 00:07:52.470689 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:52Z","lastTransitionTime":"2025-12-03T00:07:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:52 crc kubenswrapper[4953]: I1203 00:07:52.573822 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:52 crc kubenswrapper[4953]: I1203 00:07:52.573926 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:52 crc kubenswrapper[4953]: I1203 00:07:52.573942 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:52 crc kubenswrapper[4953]: I1203 00:07:52.573966 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:52 crc kubenswrapper[4953]: I1203 00:07:52.574016 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:52Z","lastTransitionTime":"2025-12-03T00:07:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:52 crc kubenswrapper[4953]: I1203 00:07:52.677118 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:52 crc kubenswrapper[4953]: I1203 00:07:52.677194 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:52 crc kubenswrapper[4953]: I1203 00:07:52.677216 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:52 crc kubenswrapper[4953]: I1203 00:07:52.677246 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:52 crc kubenswrapper[4953]: I1203 00:07:52.677268 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:52Z","lastTransitionTime":"2025-12-03T00:07:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:52 crc kubenswrapper[4953]: I1203 00:07:52.780530 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:52 crc kubenswrapper[4953]: I1203 00:07:52.780612 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:52 crc kubenswrapper[4953]: I1203 00:07:52.780639 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:52 crc kubenswrapper[4953]: I1203 00:07:52.780672 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:52 crc kubenswrapper[4953]: I1203 00:07:52.780693 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:52Z","lastTransitionTime":"2025-12-03T00:07:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:52 crc kubenswrapper[4953]: I1203 00:07:52.883960 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:52 crc kubenswrapper[4953]: I1203 00:07:52.884069 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:52 crc kubenswrapper[4953]: I1203 00:07:52.884088 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:52 crc kubenswrapper[4953]: I1203 00:07:52.884112 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:52 crc kubenswrapper[4953]: I1203 00:07:52.884129 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:52Z","lastTransitionTime":"2025-12-03T00:07:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:52 crc kubenswrapper[4953]: I1203 00:07:52.986933 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:52 crc kubenswrapper[4953]: I1203 00:07:52.987086 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:52 crc kubenswrapper[4953]: I1203 00:07:52.987113 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:52 crc kubenswrapper[4953]: I1203 00:07:52.987145 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:52 crc kubenswrapper[4953]: I1203 00:07:52.987168 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:52Z","lastTransitionTime":"2025-12-03T00:07:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:53 crc kubenswrapper[4953]: I1203 00:07:53.090017 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:53 crc kubenswrapper[4953]: I1203 00:07:53.090069 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:53 crc kubenswrapper[4953]: I1203 00:07:53.090084 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:53 crc kubenswrapper[4953]: I1203 00:07:53.090100 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:53 crc kubenswrapper[4953]: I1203 00:07:53.090113 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:53Z","lastTransitionTime":"2025-12-03T00:07:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:53 crc kubenswrapper[4953]: I1203 00:07:53.193047 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:53 crc kubenswrapper[4953]: I1203 00:07:53.193100 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:53 crc kubenswrapper[4953]: I1203 00:07:53.193116 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:53 crc kubenswrapper[4953]: I1203 00:07:53.193139 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:53 crc kubenswrapper[4953]: I1203 00:07:53.193158 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:53Z","lastTransitionTime":"2025-12-03T00:07:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:53 crc kubenswrapper[4953]: I1203 00:07:53.295737 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:53 crc kubenswrapper[4953]: I1203 00:07:53.296271 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:53 crc kubenswrapper[4953]: I1203 00:07:53.296480 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:53 crc kubenswrapper[4953]: I1203 00:07:53.296759 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:53 crc kubenswrapper[4953]: I1203 00:07:53.296913 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:53Z","lastTransitionTime":"2025-12-03T00:07:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:53 crc kubenswrapper[4953]: I1203 00:07:53.348221 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xshk4" Dec 03 00:07:53 crc kubenswrapper[4953]: E1203 00:07:53.348405 4953 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xshk4" podUID="bd03a758-0ca9-47d8-86f7-18b2d610df43" Dec 03 00:07:53 crc kubenswrapper[4953]: I1203 00:07:53.402923 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:53 crc kubenswrapper[4953]: I1203 00:07:53.403008 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:53 crc kubenswrapper[4953]: I1203 00:07:53.403032 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:53 crc kubenswrapper[4953]: I1203 00:07:53.403058 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:53 crc kubenswrapper[4953]: I1203 00:07:53.403080 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:53Z","lastTransitionTime":"2025-12-03T00:07:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:53 crc kubenswrapper[4953]: I1203 00:07:53.505563 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:53 crc kubenswrapper[4953]: I1203 00:07:53.505610 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:53 crc kubenswrapper[4953]: I1203 00:07:53.505631 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:53 crc kubenswrapper[4953]: I1203 00:07:53.505659 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:53 crc kubenswrapper[4953]: I1203 00:07:53.505680 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:53Z","lastTransitionTime":"2025-12-03T00:07:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:53 crc kubenswrapper[4953]: I1203 00:07:53.608683 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:53 crc kubenswrapper[4953]: I1203 00:07:53.608774 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:53 crc kubenswrapper[4953]: I1203 00:07:53.608814 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:53 crc kubenswrapper[4953]: I1203 00:07:53.608846 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:53 crc kubenswrapper[4953]: I1203 00:07:53.608867 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:53Z","lastTransitionTime":"2025-12-03T00:07:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:53 crc kubenswrapper[4953]: I1203 00:07:53.712037 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:53 crc kubenswrapper[4953]: I1203 00:07:53.712102 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:53 crc kubenswrapper[4953]: I1203 00:07:53.712124 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:53 crc kubenswrapper[4953]: I1203 00:07:53.712178 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:53 crc kubenswrapper[4953]: I1203 00:07:53.712198 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:53Z","lastTransitionTime":"2025-12-03T00:07:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:53 crc kubenswrapper[4953]: I1203 00:07:53.814633 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:53 crc kubenswrapper[4953]: I1203 00:07:53.814704 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:53 crc kubenswrapper[4953]: I1203 00:07:53.814719 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:53 crc kubenswrapper[4953]: I1203 00:07:53.814746 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:53 crc kubenswrapper[4953]: I1203 00:07:53.814765 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:53Z","lastTransitionTime":"2025-12-03T00:07:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:53 crc kubenswrapper[4953]: I1203 00:07:53.816405 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:53 crc kubenswrapper[4953]: I1203 00:07:53.816463 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:53 crc kubenswrapper[4953]: I1203 00:07:53.816480 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:53 crc kubenswrapper[4953]: I1203 00:07:53.816506 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:53 crc kubenswrapper[4953]: I1203 00:07:53.816524 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:53Z","lastTransitionTime":"2025-12-03T00:07:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:53 crc kubenswrapper[4953]: E1203 00:07:53.837231 4953 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:07:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:53Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:07:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:53Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:07:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:53Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:07:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:53Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"eb4ae5a6-0f6f-49ef-a027-1cdee9d3c536\\\",\\\"systemUUID\\\":\\\"4c1f3d44-3d3d-425d-993f-471223206a89\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:53Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:53 crc kubenswrapper[4953]: I1203 00:07:53.842672 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:53 crc kubenswrapper[4953]: I1203 00:07:53.842894 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:53 crc kubenswrapper[4953]: I1203 00:07:53.843109 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:53 crc kubenswrapper[4953]: I1203 00:07:53.843344 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:53 crc kubenswrapper[4953]: I1203 00:07:53.843551 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:53Z","lastTransitionTime":"2025-12-03T00:07:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:53 crc kubenswrapper[4953]: E1203 00:07:53.856910 4953 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:07:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:53Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:07:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:53Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:07:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:53Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:07:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:53Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"eb4ae5a6-0f6f-49ef-a027-1cdee9d3c536\\\",\\\"systemUUID\\\":\\\"4c1f3d44-3d3d-425d-993f-471223206a89\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:53Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:53 crc kubenswrapper[4953]: I1203 00:07:53.861789 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:53 crc kubenswrapper[4953]: I1203 00:07:53.862066 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:53 crc kubenswrapper[4953]: I1203 00:07:53.862232 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:53 crc kubenswrapper[4953]: I1203 00:07:53.862409 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:53 crc kubenswrapper[4953]: I1203 00:07:53.862602 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:53Z","lastTransitionTime":"2025-12-03T00:07:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:53 crc kubenswrapper[4953]: E1203 00:07:53.876772 4953 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:07:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:53Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:07:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:53Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:07:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:53Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:07:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:53Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"eb4ae5a6-0f6f-49ef-a027-1cdee9d3c536\\\",\\\"systemUUID\\\":\\\"4c1f3d44-3d3d-425d-993f-471223206a89\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:53Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:53 crc kubenswrapper[4953]: I1203 00:07:53.880618 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:53 crc kubenswrapper[4953]: I1203 00:07:53.880664 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:53 crc kubenswrapper[4953]: I1203 00:07:53.880682 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:53 crc kubenswrapper[4953]: I1203 00:07:53.880704 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:53 crc kubenswrapper[4953]: I1203 00:07:53.880724 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:53Z","lastTransitionTime":"2025-12-03T00:07:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:53 crc kubenswrapper[4953]: E1203 00:07:53.895680 4953 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:07:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:53Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:07:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:53Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:07:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:53Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:07:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:53Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"eb4ae5a6-0f6f-49ef-a027-1cdee9d3c536\\\",\\\"systemUUID\\\":\\\"4c1f3d44-3d3d-425d-993f-471223206a89\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:53Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:53 crc kubenswrapper[4953]: I1203 00:07:53.899680 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:53 crc kubenswrapper[4953]: I1203 00:07:53.899822 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:53 crc kubenswrapper[4953]: I1203 00:07:53.899913 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:53 crc kubenswrapper[4953]: I1203 00:07:53.901600 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:53 crc kubenswrapper[4953]: I1203 00:07:53.901720 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:53Z","lastTransitionTime":"2025-12-03T00:07:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:53 crc kubenswrapper[4953]: E1203 00:07:53.914656 4953 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:07:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:53Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:07:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:53Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:07:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:53Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:07:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:53Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"eb4ae5a6-0f6f-49ef-a027-1cdee9d3c536\\\",\\\"systemUUID\\\":\\\"4c1f3d44-3d3d-425d-993f-471223206a89\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:07:53Z is after 2025-08-24T17:21:41Z" Dec 03 00:07:53 crc kubenswrapper[4953]: E1203 00:07:53.914886 4953 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 03 00:07:53 crc kubenswrapper[4953]: I1203 00:07:53.917501 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:53 crc kubenswrapper[4953]: I1203 00:07:53.917556 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:53 crc kubenswrapper[4953]: I1203 00:07:53.917576 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:53 crc kubenswrapper[4953]: I1203 00:07:53.917595 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:53 crc kubenswrapper[4953]: I1203 00:07:53.917608 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:53Z","lastTransitionTime":"2025-12-03T00:07:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:54 crc kubenswrapper[4953]: I1203 00:07:54.020092 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:54 crc kubenswrapper[4953]: I1203 00:07:54.020472 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:54 crc kubenswrapper[4953]: I1203 00:07:54.020608 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:54 crc kubenswrapper[4953]: I1203 00:07:54.020803 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:54 crc kubenswrapper[4953]: I1203 00:07:54.021042 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:54Z","lastTransitionTime":"2025-12-03T00:07:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:54 crc kubenswrapper[4953]: I1203 00:07:54.124490 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:54 crc kubenswrapper[4953]: I1203 00:07:54.124575 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:54 crc kubenswrapper[4953]: I1203 00:07:54.124592 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:54 crc kubenswrapper[4953]: I1203 00:07:54.124650 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:54 crc kubenswrapper[4953]: I1203 00:07:54.124668 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:54Z","lastTransitionTime":"2025-12-03T00:07:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:54 crc kubenswrapper[4953]: I1203 00:07:54.228685 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:54 crc kubenswrapper[4953]: I1203 00:07:54.228781 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:54 crc kubenswrapper[4953]: I1203 00:07:54.228803 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:54 crc kubenswrapper[4953]: I1203 00:07:54.229012 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:54 crc kubenswrapper[4953]: I1203 00:07:54.229037 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:54Z","lastTransitionTime":"2025-12-03T00:07:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:54 crc kubenswrapper[4953]: I1203 00:07:54.332186 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:54 crc kubenswrapper[4953]: I1203 00:07:54.332330 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:54 crc kubenswrapper[4953]: I1203 00:07:54.332356 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:54 crc kubenswrapper[4953]: I1203 00:07:54.332384 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:54 crc kubenswrapper[4953]: I1203 00:07:54.332474 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:54Z","lastTransitionTime":"2025-12-03T00:07:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:54 crc kubenswrapper[4953]: I1203 00:07:54.347117 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 00:07:54 crc kubenswrapper[4953]: I1203 00:07:54.347136 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 00:07:54 crc kubenswrapper[4953]: E1203 00:07:54.347254 4953 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 00:07:54 crc kubenswrapper[4953]: I1203 00:07:54.347299 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 00:07:54 crc kubenswrapper[4953]: E1203 00:07:54.347413 4953 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 00:07:54 crc kubenswrapper[4953]: E1203 00:07:54.347553 4953 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 00:07:54 crc kubenswrapper[4953]: I1203 00:07:54.435510 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:54 crc kubenswrapper[4953]: I1203 00:07:54.435621 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:54 crc kubenswrapper[4953]: I1203 00:07:54.435656 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:54 crc kubenswrapper[4953]: I1203 00:07:54.435745 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:54 crc kubenswrapper[4953]: I1203 00:07:54.435771 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:54Z","lastTransitionTime":"2025-12-03T00:07:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:54 crc kubenswrapper[4953]: I1203 00:07:54.542636 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:54 crc kubenswrapper[4953]: I1203 00:07:54.542926 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:54 crc kubenswrapper[4953]: I1203 00:07:54.543211 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:54 crc kubenswrapper[4953]: I1203 00:07:54.543245 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:54 crc kubenswrapper[4953]: I1203 00:07:54.543270 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:54Z","lastTransitionTime":"2025-12-03T00:07:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:54 crc kubenswrapper[4953]: I1203 00:07:54.645594 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:54 crc kubenswrapper[4953]: I1203 00:07:54.645641 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:54 crc kubenswrapper[4953]: I1203 00:07:54.645679 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:54 crc kubenswrapper[4953]: I1203 00:07:54.645701 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:54 crc kubenswrapper[4953]: I1203 00:07:54.645715 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:54Z","lastTransitionTime":"2025-12-03T00:07:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:54 crc kubenswrapper[4953]: I1203 00:07:54.749446 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:54 crc kubenswrapper[4953]: I1203 00:07:54.749519 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:54 crc kubenswrapper[4953]: I1203 00:07:54.749542 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:54 crc kubenswrapper[4953]: I1203 00:07:54.749569 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:54 crc kubenswrapper[4953]: I1203 00:07:54.749589 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:54Z","lastTransitionTime":"2025-12-03T00:07:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:54 crc kubenswrapper[4953]: I1203 00:07:54.851379 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:54 crc kubenswrapper[4953]: I1203 00:07:54.851414 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:54 crc kubenswrapper[4953]: I1203 00:07:54.851424 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:54 crc kubenswrapper[4953]: I1203 00:07:54.851439 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:54 crc kubenswrapper[4953]: I1203 00:07:54.851453 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:54Z","lastTransitionTime":"2025-12-03T00:07:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:54 crc kubenswrapper[4953]: I1203 00:07:54.955338 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:54 crc kubenswrapper[4953]: I1203 00:07:54.955663 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:54 crc kubenswrapper[4953]: I1203 00:07:54.955746 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:54 crc kubenswrapper[4953]: I1203 00:07:54.955873 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:54 crc kubenswrapper[4953]: I1203 00:07:54.956032 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:54Z","lastTransitionTime":"2025-12-03T00:07:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:55 crc kubenswrapper[4953]: I1203 00:07:55.058944 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:55 crc kubenswrapper[4953]: I1203 00:07:55.059005 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:55 crc kubenswrapper[4953]: I1203 00:07:55.059017 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:55 crc kubenswrapper[4953]: I1203 00:07:55.059033 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:55 crc kubenswrapper[4953]: I1203 00:07:55.059043 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:55Z","lastTransitionTime":"2025-12-03T00:07:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:55 crc kubenswrapper[4953]: I1203 00:07:55.161931 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:55 crc kubenswrapper[4953]: I1203 00:07:55.161963 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:55 crc kubenswrapper[4953]: I1203 00:07:55.161987 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:55 crc kubenswrapper[4953]: I1203 00:07:55.162002 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:55 crc kubenswrapper[4953]: I1203 00:07:55.162012 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:55Z","lastTransitionTime":"2025-12-03T00:07:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:55 crc kubenswrapper[4953]: I1203 00:07:55.265020 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:55 crc kubenswrapper[4953]: I1203 00:07:55.265509 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:55 crc kubenswrapper[4953]: I1203 00:07:55.265613 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:55 crc kubenswrapper[4953]: I1203 00:07:55.265739 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:55 crc kubenswrapper[4953]: I1203 00:07:55.265854 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:55Z","lastTransitionTime":"2025-12-03T00:07:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:55 crc kubenswrapper[4953]: I1203 00:07:55.347381 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xshk4" Dec 03 00:07:55 crc kubenswrapper[4953]: E1203 00:07:55.347528 4953 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xshk4" podUID="bd03a758-0ca9-47d8-86f7-18b2d610df43" Dec 03 00:07:55 crc kubenswrapper[4953]: I1203 00:07:55.367790 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:55 crc kubenswrapper[4953]: I1203 00:07:55.367826 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:55 crc kubenswrapper[4953]: I1203 00:07:55.367836 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:55 crc kubenswrapper[4953]: I1203 00:07:55.367847 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:55 crc kubenswrapper[4953]: I1203 00:07:55.367857 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:55Z","lastTransitionTime":"2025-12-03T00:07:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:55 crc kubenswrapper[4953]: I1203 00:07:55.470861 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:55 crc kubenswrapper[4953]: I1203 00:07:55.470899 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:55 crc kubenswrapper[4953]: I1203 00:07:55.470911 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:55 crc kubenswrapper[4953]: I1203 00:07:55.470925 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:55 crc kubenswrapper[4953]: I1203 00:07:55.470935 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:55Z","lastTransitionTime":"2025-12-03T00:07:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:55 crc kubenswrapper[4953]: I1203 00:07:55.574228 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:55 crc kubenswrapper[4953]: I1203 00:07:55.574281 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:55 crc kubenswrapper[4953]: I1203 00:07:55.574294 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:55 crc kubenswrapper[4953]: I1203 00:07:55.574311 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:55 crc kubenswrapper[4953]: I1203 00:07:55.574324 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:55Z","lastTransitionTime":"2025-12-03T00:07:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:55 crc kubenswrapper[4953]: I1203 00:07:55.679308 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:55 crc kubenswrapper[4953]: I1203 00:07:55.679362 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:55 crc kubenswrapper[4953]: I1203 00:07:55.679373 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:55 crc kubenswrapper[4953]: I1203 00:07:55.679393 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:55 crc kubenswrapper[4953]: I1203 00:07:55.679405 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:55Z","lastTransitionTime":"2025-12-03T00:07:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:55 crc kubenswrapper[4953]: I1203 00:07:55.782290 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:55 crc kubenswrapper[4953]: I1203 00:07:55.782362 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:55 crc kubenswrapper[4953]: I1203 00:07:55.782385 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:55 crc kubenswrapper[4953]: I1203 00:07:55.782418 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:55 crc kubenswrapper[4953]: I1203 00:07:55.782442 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:55Z","lastTransitionTime":"2025-12-03T00:07:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:55 crc kubenswrapper[4953]: I1203 00:07:55.886217 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:55 crc kubenswrapper[4953]: I1203 00:07:55.886280 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:55 crc kubenswrapper[4953]: I1203 00:07:55.886291 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:55 crc kubenswrapper[4953]: I1203 00:07:55.886314 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:55 crc kubenswrapper[4953]: I1203 00:07:55.886328 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:55Z","lastTransitionTime":"2025-12-03T00:07:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:55 crc kubenswrapper[4953]: I1203 00:07:55.989285 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:55 crc kubenswrapper[4953]: I1203 00:07:55.989344 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:55 crc kubenswrapper[4953]: I1203 00:07:55.989362 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:55 crc kubenswrapper[4953]: I1203 00:07:55.989386 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:55 crc kubenswrapper[4953]: I1203 00:07:55.989402 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:55Z","lastTransitionTime":"2025-12-03T00:07:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:56 crc kubenswrapper[4953]: I1203 00:07:56.092047 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:56 crc kubenswrapper[4953]: I1203 00:07:56.092091 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:56 crc kubenswrapper[4953]: I1203 00:07:56.092100 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:56 crc kubenswrapper[4953]: I1203 00:07:56.092131 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:56 crc kubenswrapper[4953]: I1203 00:07:56.092141 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:56Z","lastTransitionTime":"2025-12-03T00:07:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:56 crc kubenswrapper[4953]: I1203 00:07:56.194677 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:56 crc kubenswrapper[4953]: I1203 00:07:56.194751 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:56 crc kubenswrapper[4953]: I1203 00:07:56.194772 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:56 crc kubenswrapper[4953]: I1203 00:07:56.194799 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:56 crc kubenswrapper[4953]: I1203 00:07:56.194818 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:56Z","lastTransitionTime":"2025-12-03T00:07:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:56 crc kubenswrapper[4953]: I1203 00:07:56.297184 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:56 crc kubenswrapper[4953]: I1203 00:07:56.297240 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:56 crc kubenswrapper[4953]: I1203 00:07:56.297254 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:56 crc kubenswrapper[4953]: I1203 00:07:56.297279 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:56 crc kubenswrapper[4953]: I1203 00:07:56.297293 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:56Z","lastTransitionTime":"2025-12-03T00:07:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:56 crc kubenswrapper[4953]: I1203 00:07:56.347668 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 00:07:56 crc kubenswrapper[4953]: I1203 00:07:56.347700 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 00:07:56 crc kubenswrapper[4953]: I1203 00:07:56.347664 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 00:07:56 crc kubenswrapper[4953]: E1203 00:07:56.347815 4953 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 00:07:56 crc kubenswrapper[4953]: E1203 00:07:56.347887 4953 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 00:07:56 crc kubenswrapper[4953]: E1203 00:07:56.348027 4953 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 00:07:56 crc kubenswrapper[4953]: I1203 00:07:56.400669 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:56 crc kubenswrapper[4953]: I1203 00:07:56.400734 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:56 crc kubenswrapper[4953]: I1203 00:07:56.400756 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:56 crc kubenswrapper[4953]: I1203 00:07:56.400785 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:56 crc kubenswrapper[4953]: I1203 00:07:56.400806 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:56Z","lastTransitionTime":"2025-12-03T00:07:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:56 crc kubenswrapper[4953]: I1203 00:07:56.504331 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:56 crc kubenswrapper[4953]: I1203 00:07:56.504374 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:56 crc kubenswrapper[4953]: I1203 00:07:56.504384 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:56 crc kubenswrapper[4953]: I1203 00:07:56.504403 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:56 crc kubenswrapper[4953]: I1203 00:07:56.504416 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:56Z","lastTransitionTime":"2025-12-03T00:07:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:56 crc kubenswrapper[4953]: I1203 00:07:56.607653 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:56 crc kubenswrapper[4953]: I1203 00:07:56.607705 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:56 crc kubenswrapper[4953]: I1203 00:07:56.607714 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:56 crc kubenswrapper[4953]: I1203 00:07:56.607732 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:56 crc kubenswrapper[4953]: I1203 00:07:56.607743 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:56Z","lastTransitionTime":"2025-12-03T00:07:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:56 crc kubenswrapper[4953]: I1203 00:07:56.711118 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:56 crc kubenswrapper[4953]: I1203 00:07:56.711168 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:56 crc kubenswrapper[4953]: I1203 00:07:56.711177 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:56 crc kubenswrapper[4953]: I1203 00:07:56.711199 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:56 crc kubenswrapper[4953]: I1203 00:07:56.711212 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:56Z","lastTransitionTime":"2025-12-03T00:07:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:56 crc kubenswrapper[4953]: I1203 00:07:56.814341 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:56 crc kubenswrapper[4953]: I1203 00:07:56.814385 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:56 crc kubenswrapper[4953]: I1203 00:07:56.814396 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:56 crc kubenswrapper[4953]: I1203 00:07:56.814436 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:56 crc kubenswrapper[4953]: I1203 00:07:56.814447 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:56Z","lastTransitionTime":"2025-12-03T00:07:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:56 crc kubenswrapper[4953]: I1203 00:07:56.917189 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:56 crc kubenswrapper[4953]: I1203 00:07:56.917230 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:56 crc kubenswrapper[4953]: I1203 00:07:56.917239 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:56 crc kubenswrapper[4953]: I1203 00:07:56.917252 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:56 crc kubenswrapper[4953]: I1203 00:07:56.917262 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:56Z","lastTransitionTime":"2025-12-03T00:07:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:57 crc kubenswrapper[4953]: I1203 00:07:57.020007 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:57 crc kubenswrapper[4953]: I1203 00:07:57.020046 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:57 crc kubenswrapper[4953]: I1203 00:07:57.020057 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:57 crc kubenswrapper[4953]: I1203 00:07:57.020072 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:57 crc kubenswrapper[4953]: I1203 00:07:57.020084 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:57Z","lastTransitionTime":"2025-12-03T00:07:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:57 crc kubenswrapper[4953]: I1203 00:07:57.123147 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:57 crc kubenswrapper[4953]: I1203 00:07:57.123193 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:57 crc kubenswrapper[4953]: I1203 00:07:57.123203 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:57 crc kubenswrapper[4953]: I1203 00:07:57.123218 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:57 crc kubenswrapper[4953]: I1203 00:07:57.123228 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:57Z","lastTransitionTime":"2025-12-03T00:07:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:57 crc kubenswrapper[4953]: I1203 00:07:57.226710 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:57 crc kubenswrapper[4953]: I1203 00:07:57.226775 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:57 crc kubenswrapper[4953]: I1203 00:07:57.226799 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:57 crc kubenswrapper[4953]: I1203 00:07:57.226827 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:57 crc kubenswrapper[4953]: I1203 00:07:57.226852 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:57Z","lastTransitionTime":"2025-12-03T00:07:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:57 crc kubenswrapper[4953]: I1203 00:07:57.333933 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:57 crc kubenswrapper[4953]: I1203 00:07:57.334232 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:57 crc kubenswrapper[4953]: I1203 00:07:57.334248 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:57 crc kubenswrapper[4953]: I1203 00:07:57.334269 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:57 crc kubenswrapper[4953]: I1203 00:07:57.334285 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:57Z","lastTransitionTime":"2025-12-03T00:07:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:57 crc kubenswrapper[4953]: I1203 00:07:57.347596 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xshk4" Dec 03 00:07:57 crc kubenswrapper[4953]: E1203 00:07:57.347827 4953 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xshk4" podUID="bd03a758-0ca9-47d8-86f7-18b2d610df43" Dec 03 00:07:57 crc kubenswrapper[4953]: I1203 00:07:57.437294 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:57 crc kubenswrapper[4953]: I1203 00:07:57.437328 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:57 crc kubenswrapper[4953]: I1203 00:07:57.437358 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:57 crc kubenswrapper[4953]: I1203 00:07:57.437372 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:57 crc kubenswrapper[4953]: I1203 00:07:57.437381 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:57Z","lastTransitionTime":"2025-12-03T00:07:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:57 crc kubenswrapper[4953]: I1203 00:07:57.541438 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:57 crc kubenswrapper[4953]: I1203 00:07:57.541512 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:57 crc kubenswrapper[4953]: I1203 00:07:57.541523 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:57 crc kubenswrapper[4953]: I1203 00:07:57.541544 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:57 crc kubenswrapper[4953]: I1203 00:07:57.541558 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:57Z","lastTransitionTime":"2025-12-03T00:07:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:57 crc kubenswrapper[4953]: I1203 00:07:57.644903 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:57 crc kubenswrapper[4953]: I1203 00:07:57.644995 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:57 crc kubenswrapper[4953]: I1203 00:07:57.645012 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:57 crc kubenswrapper[4953]: I1203 00:07:57.645038 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:57 crc kubenswrapper[4953]: I1203 00:07:57.645052 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:57Z","lastTransitionTime":"2025-12-03T00:07:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:57 crc kubenswrapper[4953]: I1203 00:07:57.747894 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:57 crc kubenswrapper[4953]: I1203 00:07:57.747941 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:57 crc kubenswrapper[4953]: I1203 00:07:57.747954 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:57 crc kubenswrapper[4953]: I1203 00:07:57.748000 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:57 crc kubenswrapper[4953]: I1203 00:07:57.748014 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:57Z","lastTransitionTime":"2025-12-03T00:07:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:57 crc kubenswrapper[4953]: I1203 00:07:57.851027 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:57 crc kubenswrapper[4953]: I1203 00:07:57.851089 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:57 crc kubenswrapper[4953]: I1203 00:07:57.851107 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:57 crc kubenswrapper[4953]: I1203 00:07:57.851133 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:57 crc kubenswrapper[4953]: I1203 00:07:57.851152 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:57Z","lastTransitionTime":"2025-12-03T00:07:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:57 crc kubenswrapper[4953]: I1203 00:07:57.954826 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:57 crc kubenswrapper[4953]: I1203 00:07:57.954964 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:57 crc kubenswrapper[4953]: I1203 00:07:57.955034 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:57 crc kubenswrapper[4953]: I1203 00:07:57.955063 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:57 crc kubenswrapper[4953]: I1203 00:07:57.955122 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:57Z","lastTransitionTime":"2025-12-03T00:07:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:58 crc kubenswrapper[4953]: I1203 00:07:58.057943 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:58 crc kubenswrapper[4953]: I1203 00:07:58.058039 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:58 crc kubenswrapper[4953]: I1203 00:07:58.058058 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:58 crc kubenswrapper[4953]: I1203 00:07:58.058081 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:58 crc kubenswrapper[4953]: I1203 00:07:58.058100 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:58Z","lastTransitionTime":"2025-12-03T00:07:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:58 crc kubenswrapper[4953]: I1203 00:07:58.161065 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:58 crc kubenswrapper[4953]: I1203 00:07:58.161123 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:58 crc kubenswrapper[4953]: I1203 00:07:58.161138 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:58 crc kubenswrapper[4953]: I1203 00:07:58.161163 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:58 crc kubenswrapper[4953]: I1203 00:07:58.161178 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:58Z","lastTransitionTime":"2025-12-03T00:07:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:58 crc kubenswrapper[4953]: I1203 00:07:58.264712 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:58 crc kubenswrapper[4953]: I1203 00:07:58.264765 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:58 crc kubenswrapper[4953]: I1203 00:07:58.264777 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:58 crc kubenswrapper[4953]: I1203 00:07:58.264801 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:58 crc kubenswrapper[4953]: I1203 00:07:58.264814 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:58Z","lastTransitionTime":"2025-12-03T00:07:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:58 crc kubenswrapper[4953]: I1203 00:07:58.347330 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 00:07:58 crc kubenswrapper[4953]: I1203 00:07:58.347411 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 00:07:58 crc kubenswrapper[4953]: E1203 00:07:58.347473 4953 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 00:07:58 crc kubenswrapper[4953]: I1203 00:07:58.347331 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 00:07:58 crc kubenswrapper[4953]: E1203 00:07:58.347603 4953 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 00:07:58 crc kubenswrapper[4953]: E1203 00:07:58.347839 4953 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 00:07:58 crc kubenswrapper[4953]: I1203 00:07:58.367694 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:58 crc kubenswrapper[4953]: I1203 00:07:58.367741 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:58 crc kubenswrapper[4953]: I1203 00:07:58.367754 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:58 crc kubenswrapper[4953]: I1203 00:07:58.367780 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:58 crc kubenswrapper[4953]: I1203 00:07:58.367805 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:58Z","lastTransitionTime":"2025-12-03T00:07:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:58 crc kubenswrapper[4953]: I1203 00:07:58.471408 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:58 crc kubenswrapper[4953]: I1203 00:07:58.471480 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:58 crc kubenswrapper[4953]: I1203 00:07:58.471494 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:58 crc kubenswrapper[4953]: I1203 00:07:58.471516 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:58 crc kubenswrapper[4953]: I1203 00:07:58.471528 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:58Z","lastTransitionTime":"2025-12-03T00:07:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:58 crc kubenswrapper[4953]: I1203 00:07:58.575174 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:58 crc kubenswrapper[4953]: I1203 00:07:58.575253 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:58 crc kubenswrapper[4953]: I1203 00:07:58.575271 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:58 crc kubenswrapper[4953]: I1203 00:07:58.575302 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:58 crc kubenswrapper[4953]: I1203 00:07:58.575323 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:58Z","lastTransitionTime":"2025-12-03T00:07:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:58 crc kubenswrapper[4953]: I1203 00:07:58.678222 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:58 crc kubenswrapper[4953]: I1203 00:07:58.678301 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:58 crc kubenswrapper[4953]: I1203 00:07:58.678315 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:58 crc kubenswrapper[4953]: I1203 00:07:58.678344 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:58 crc kubenswrapper[4953]: I1203 00:07:58.678359 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:58Z","lastTransitionTime":"2025-12-03T00:07:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:58 crc kubenswrapper[4953]: I1203 00:07:58.781590 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:58 crc kubenswrapper[4953]: I1203 00:07:58.781629 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:58 crc kubenswrapper[4953]: I1203 00:07:58.781637 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:58 crc kubenswrapper[4953]: I1203 00:07:58.781656 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:58 crc kubenswrapper[4953]: I1203 00:07:58.781669 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:58Z","lastTransitionTime":"2025-12-03T00:07:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:58 crc kubenswrapper[4953]: I1203 00:07:58.884672 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:58 crc kubenswrapper[4953]: I1203 00:07:58.884705 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:58 crc kubenswrapper[4953]: I1203 00:07:58.884714 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:58 crc kubenswrapper[4953]: I1203 00:07:58.884728 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:58 crc kubenswrapper[4953]: I1203 00:07:58.884740 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:58Z","lastTransitionTime":"2025-12-03T00:07:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:58 crc kubenswrapper[4953]: I1203 00:07:58.987771 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:58 crc kubenswrapper[4953]: I1203 00:07:58.987822 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:58 crc kubenswrapper[4953]: I1203 00:07:58.987833 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:58 crc kubenswrapper[4953]: I1203 00:07:58.987847 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:58 crc kubenswrapper[4953]: I1203 00:07:58.987855 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:58Z","lastTransitionTime":"2025-12-03T00:07:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:59 crc kubenswrapper[4953]: I1203 00:07:59.091291 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:59 crc kubenswrapper[4953]: I1203 00:07:59.091373 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:59 crc kubenswrapper[4953]: I1203 00:07:59.091400 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:59 crc kubenswrapper[4953]: I1203 00:07:59.091434 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:59 crc kubenswrapper[4953]: I1203 00:07:59.091458 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:59Z","lastTransitionTime":"2025-12-03T00:07:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:59 crc kubenswrapper[4953]: I1203 00:07:59.194209 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:59 crc kubenswrapper[4953]: I1203 00:07:59.194267 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:59 crc kubenswrapper[4953]: I1203 00:07:59.194277 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:59 crc kubenswrapper[4953]: I1203 00:07:59.194289 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:59 crc kubenswrapper[4953]: I1203 00:07:59.194298 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:59Z","lastTransitionTime":"2025-12-03T00:07:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:59 crc kubenswrapper[4953]: I1203 00:07:59.297122 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:59 crc kubenswrapper[4953]: I1203 00:07:59.297191 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:59 crc kubenswrapper[4953]: I1203 00:07:59.297209 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:59 crc kubenswrapper[4953]: I1203 00:07:59.297234 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:59 crc kubenswrapper[4953]: I1203 00:07:59.297252 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:59Z","lastTransitionTime":"2025-12-03T00:07:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:59 crc kubenswrapper[4953]: I1203 00:07:59.348006 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xshk4" Dec 03 00:07:59 crc kubenswrapper[4953]: E1203 00:07:59.348219 4953 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xshk4" podUID="bd03a758-0ca9-47d8-86f7-18b2d610df43" Dec 03 00:07:59 crc kubenswrapper[4953]: I1203 00:07:59.364278 4953 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc"] Dec 03 00:07:59 crc kubenswrapper[4953]: I1203 00:07:59.400164 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:59 crc kubenswrapper[4953]: I1203 00:07:59.400226 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:59 crc kubenswrapper[4953]: I1203 00:07:59.400247 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:59 crc kubenswrapper[4953]: I1203 00:07:59.400276 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:59 crc kubenswrapper[4953]: I1203 00:07:59.400297 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:59Z","lastTransitionTime":"2025-12-03T00:07:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:59 crc kubenswrapper[4953]: I1203 00:07:59.503105 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:59 crc kubenswrapper[4953]: I1203 00:07:59.503160 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:59 crc kubenswrapper[4953]: I1203 00:07:59.503170 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:59 crc kubenswrapper[4953]: I1203 00:07:59.503191 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:59 crc kubenswrapper[4953]: I1203 00:07:59.503206 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:59Z","lastTransitionTime":"2025-12-03T00:07:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:59 crc kubenswrapper[4953]: I1203 00:07:59.606808 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:59 crc kubenswrapper[4953]: I1203 00:07:59.606860 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:59 crc kubenswrapper[4953]: I1203 00:07:59.606872 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:59 crc kubenswrapper[4953]: I1203 00:07:59.606894 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:59 crc kubenswrapper[4953]: I1203 00:07:59.606908 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:59Z","lastTransitionTime":"2025-12-03T00:07:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:59 crc kubenswrapper[4953]: I1203 00:07:59.708781 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:59 crc kubenswrapper[4953]: I1203 00:07:59.708821 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:59 crc kubenswrapper[4953]: I1203 00:07:59.708830 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:59 crc kubenswrapper[4953]: I1203 00:07:59.708846 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:59 crc kubenswrapper[4953]: I1203 00:07:59.708857 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:59Z","lastTransitionTime":"2025-12-03T00:07:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:59 crc kubenswrapper[4953]: I1203 00:07:59.810825 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:59 crc kubenswrapper[4953]: I1203 00:07:59.810897 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:59 crc kubenswrapper[4953]: I1203 00:07:59.810913 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:59 crc kubenswrapper[4953]: I1203 00:07:59.810938 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:59 crc kubenswrapper[4953]: I1203 00:07:59.810955 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:59Z","lastTransitionTime":"2025-12-03T00:07:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:07:59 crc kubenswrapper[4953]: I1203 00:07:59.914528 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:07:59 crc kubenswrapper[4953]: I1203 00:07:59.914570 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:07:59 crc kubenswrapper[4953]: I1203 00:07:59.914583 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:07:59 crc kubenswrapper[4953]: I1203 00:07:59.914600 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:07:59 crc kubenswrapper[4953]: I1203 00:07:59.914613 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:07:59Z","lastTransitionTime":"2025-12-03T00:07:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:00 crc kubenswrapper[4953]: I1203 00:08:00.018352 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:00 crc kubenswrapper[4953]: I1203 00:08:00.018613 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:00 crc kubenswrapper[4953]: I1203 00:08:00.018682 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:00 crc kubenswrapper[4953]: I1203 00:08:00.018747 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:00 crc kubenswrapper[4953]: I1203 00:08:00.018821 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:00Z","lastTransitionTime":"2025-12-03T00:08:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:00 crc kubenswrapper[4953]: I1203 00:08:00.122123 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:00 crc kubenswrapper[4953]: I1203 00:08:00.122190 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:00 crc kubenswrapper[4953]: I1203 00:08:00.122217 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:00 crc kubenswrapper[4953]: I1203 00:08:00.122253 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:00 crc kubenswrapper[4953]: I1203 00:08:00.122279 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:00Z","lastTransitionTime":"2025-12-03T00:08:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:00 crc kubenswrapper[4953]: I1203 00:08:00.225379 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:00 crc kubenswrapper[4953]: I1203 00:08:00.225452 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:00 crc kubenswrapper[4953]: I1203 00:08:00.225463 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:00 crc kubenswrapper[4953]: I1203 00:08:00.225487 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:00 crc kubenswrapper[4953]: I1203 00:08:00.225499 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:00Z","lastTransitionTime":"2025-12-03T00:08:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:00 crc kubenswrapper[4953]: I1203 00:08:00.329515 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:00 crc kubenswrapper[4953]: I1203 00:08:00.329586 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:00 crc kubenswrapper[4953]: I1203 00:08:00.329599 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:00 crc kubenswrapper[4953]: I1203 00:08:00.329622 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:00 crc kubenswrapper[4953]: I1203 00:08:00.329637 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:00Z","lastTransitionTime":"2025-12-03T00:08:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:00 crc kubenswrapper[4953]: I1203 00:08:00.347597 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 00:08:00 crc kubenswrapper[4953]: I1203 00:08:00.347820 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 00:08:00 crc kubenswrapper[4953]: E1203 00:08:00.347844 4953 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 00:08:00 crc kubenswrapper[4953]: E1203 00:08:00.348197 4953 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 00:08:00 crc kubenswrapper[4953]: I1203 00:08:00.348248 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 00:08:00 crc kubenswrapper[4953]: E1203 00:08:00.348420 4953 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 00:08:00 crc kubenswrapper[4953]: I1203 00:08:00.433064 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:00 crc kubenswrapper[4953]: I1203 00:08:00.433123 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:00 crc kubenswrapper[4953]: I1203 00:08:00.433132 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:00 crc kubenswrapper[4953]: I1203 00:08:00.433151 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:00 crc kubenswrapper[4953]: I1203 00:08:00.433162 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:00Z","lastTransitionTime":"2025-12-03T00:08:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:00 crc kubenswrapper[4953]: I1203 00:08:00.536933 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:00 crc kubenswrapper[4953]: I1203 00:08:00.537253 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:00 crc kubenswrapper[4953]: I1203 00:08:00.537341 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:00 crc kubenswrapper[4953]: I1203 00:08:00.537431 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:00 crc kubenswrapper[4953]: I1203 00:08:00.537514 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:00Z","lastTransitionTime":"2025-12-03T00:08:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:00 crc kubenswrapper[4953]: I1203 00:08:00.640213 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:00 crc kubenswrapper[4953]: I1203 00:08:00.640271 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:00 crc kubenswrapper[4953]: I1203 00:08:00.640288 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:00 crc kubenswrapper[4953]: I1203 00:08:00.640314 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:00 crc kubenswrapper[4953]: I1203 00:08:00.640332 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:00Z","lastTransitionTime":"2025-12-03T00:08:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:00 crc kubenswrapper[4953]: I1203 00:08:00.742529 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:00 crc kubenswrapper[4953]: I1203 00:08:00.742567 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:00 crc kubenswrapper[4953]: I1203 00:08:00.742578 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:00 crc kubenswrapper[4953]: I1203 00:08:00.742593 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:00 crc kubenswrapper[4953]: I1203 00:08:00.742603 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:00Z","lastTransitionTime":"2025-12-03T00:08:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:00 crc kubenswrapper[4953]: I1203 00:08:00.844882 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:00 crc kubenswrapper[4953]: I1203 00:08:00.845173 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:00 crc kubenswrapper[4953]: I1203 00:08:00.845261 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:00 crc kubenswrapper[4953]: I1203 00:08:00.845361 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:00 crc kubenswrapper[4953]: I1203 00:08:00.845441 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:00Z","lastTransitionTime":"2025-12-03T00:08:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:00 crc kubenswrapper[4953]: I1203 00:08:00.949602 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:00 crc kubenswrapper[4953]: I1203 00:08:00.950562 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:00 crc kubenswrapper[4953]: I1203 00:08:00.950671 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:00 crc kubenswrapper[4953]: I1203 00:08:00.950786 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:00 crc kubenswrapper[4953]: I1203 00:08:00.950883 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:00Z","lastTransitionTime":"2025-12-03T00:08:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:01 crc kubenswrapper[4953]: I1203 00:08:01.054601 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:01 crc kubenswrapper[4953]: I1203 00:08:01.054652 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:01 crc kubenswrapper[4953]: I1203 00:08:01.054667 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:01 crc kubenswrapper[4953]: I1203 00:08:01.054690 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:01 crc kubenswrapper[4953]: I1203 00:08:01.054704 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:01Z","lastTransitionTime":"2025-12-03T00:08:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:01 crc kubenswrapper[4953]: I1203 00:08:01.158161 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:01 crc kubenswrapper[4953]: I1203 00:08:01.158209 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:01 crc kubenswrapper[4953]: I1203 00:08:01.158222 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:01 crc kubenswrapper[4953]: I1203 00:08:01.158242 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:01 crc kubenswrapper[4953]: I1203 00:08:01.158254 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:01Z","lastTransitionTime":"2025-12-03T00:08:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:01 crc kubenswrapper[4953]: I1203 00:08:01.260687 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:01 crc kubenswrapper[4953]: I1203 00:08:01.260737 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:01 crc kubenswrapper[4953]: I1203 00:08:01.260751 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:01 crc kubenswrapper[4953]: I1203 00:08:01.260773 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:01 crc kubenswrapper[4953]: I1203 00:08:01.260788 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:01Z","lastTransitionTime":"2025-12-03T00:08:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:01 crc kubenswrapper[4953]: I1203 00:08:01.347545 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xshk4" Dec 03 00:08:01 crc kubenswrapper[4953]: E1203 00:08:01.347784 4953 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xshk4" podUID="bd03a758-0ca9-47d8-86f7-18b2d610df43" Dec 03 00:08:01 crc kubenswrapper[4953]: I1203 00:08:01.348902 4953 scope.go:117] "RemoveContainer" containerID="dcb805f9ce285d3841e93d588fe7f3d42232382366d42b75bcfb129aefa91c1f" Dec 03 00:08:01 crc kubenswrapper[4953]: E1203 00:08:01.349217 4953 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-txffx_openshift-ovn-kubernetes(8816ebee-c735-4b8d-b1ec-65349707b347)\"" pod="openshift-ovn-kubernetes/ovnkube-node-txffx" podUID="8816ebee-c735-4b8d-b1ec-65349707b347" Dec 03 00:08:01 crc kubenswrapper[4953]: I1203 00:08:01.363641 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:10Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:08:01Z is after 2025-08-24T17:21:41Z" Dec 03 00:08:01 crc kubenswrapper[4953]: I1203 00:08:01.363868 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:01 crc kubenswrapper[4953]: I1203 00:08:01.363895 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:01 crc kubenswrapper[4953]: I1203 00:08:01.363908 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:01 crc kubenswrapper[4953]: I1203 00:08:01.363923 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:01 crc kubenswrapper[4953]: I1203 00:08:01.363934 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:01Z","lastTransitionTime":"2025-12-03T00:08:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:01 crc kubenswrapper[4953]: I1203 00:08:01.382237 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-675qk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"43a4bad8-71f4-441a-8302-aeb526a986af\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5b5794fef6477aab5deed8f642e8cc50b37eb1fd0cc6f19c003f34ad31c1e3c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gdrw4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:07:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-675qk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:08:01Z is after 2025-08-24T17:21:41Z" Dec 03 00:08:01 crc kubenswrapper[4953]: I1203 00:08:01.395589 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f94f8c51-c665-4ca3-bf17-7a04ea5c7ade\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72c5cd7dc1c3a8adfedab6ba43a59d80b3efe65f2bb56be95e76cc86a409dcf3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2810ffc1b141fa7da811406053db9d72c708edf5f2bb1dd75e4d52ab35fd24c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c67b4277b20a9f1287f50ef82d0785c9d02543a8e7ed26e0413eae332b866db7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://258d78e1ad9d3dcda6976909ae3fe71aecead5b995fc66b3fd6e2722d80b5ab8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:06:51Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:08:01Z is after 2025-08-24T17:21:41Z" Dec 03 00:08:01 crc kubenswrapper[4953]: I1203 00:08:01.406781 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5652d0f3-3fd7-4af6-8941-9119e1177641\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2030aef5b7170991268fa5dc7e7ed8ce1fbe3a2765165a10c25251426575330f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ab6052f1abf17225155ce1bd86afa742e6e78f507b7deedd7161cd8fba9120b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://650b6c79cda2e1768ba1d8763db799e210ce2748893512068d25e60f821937ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d1d60162f7f6885d4f51b4ab44163cfae6d2d04355b1dd710e26b022450005df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d1d60162f7f6885d4f51b4ab44163cfae6d2d04355b1dd710e26b022450005df\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:06:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:06:51Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:08:01Z is after 2025-08-24T17:21:41Z" Dec 03 00:08:01 crc kubenswrapper[4953]: I1203 00:08:01.420299 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09657a9f448a4c92a759c73a0a892076d7eb46393e6fd565e5f73f34c71aac23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5af5eac4a58870f63497f99b503cfa5fb922278095e0332ec50376925ecf2109\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:08:01Z is after 2025-08-24T17:21:41Z" Dec 03 00:08:01 crc kubenswrapper[4953]: I1203 00:08:01.437000 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-wj6rz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"96aa67a0-15dd-420d-a1d5-ed239d7df1fb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01815cfd3f58c3f3f5f2a894c1790e0b69c5060c48dc0b1deff050b6ad7af916\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kf62m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b12bfb5bdf9fd4f2c50a335fc82df6ee0440eaf8741897da03aa2dae368cc6b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b12bfb5bdf9fd4f2c50a335fc82df6ee0440eaf8741897da03aa2dae368cc6b4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:07:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kf62m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83aaa5fa28110bc0ef05aff75cd1442bc9bbdd33b822ae2ef22d4e4ab83238bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://83aaa5fa28110bc0ef05aff75cd1442bc9bbdd33b822ae2ef22d4e4ab83238bc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:07:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:07:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kf62m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ab6213a33a19fedebce6d74a0ebbb3b3a2de5687504e21cc24093cc42ade635\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ab6213a33a19fedebce6d74a0ebbb3b3a2de5687504e21cc24093cc42ade635\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:07:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:07:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kf62m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://de480488843afa6b8dede1e1b42b5181dca1993e2f39dc1b461a18a57e3668e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://de480488843afa6b8dede1e1b42b5181dca1993e2f39dc1b461a18a57e3668e1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:07:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kf62m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://97be4ba692bc61078d4b0ba39fb0c12772548c6c9251ff3356ea702f151223cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://97be4ba692bc61078d4b0ba39fb0c12772548c6c9251ff3356ea702f151223cd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:07:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kf62m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1a1f009fcbd8af9217db1dc97f662719f8c1db232df0e2804e44c881abb1df2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b1a1f009fcbd8af9217db1dc97f662719f8c1db232df0e2804e44c881abb1df2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:07:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:07:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kf62m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:07:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-wj6rz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:08:01Z is after 2025-08-24T17:21:41Z" Dec 03 00:08:01 crc kubenswrapper[4953]: I1203 00:08:01.448190 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-kl5s9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d85e5c97-16ad-4bd7-9e73-b7f12678be97\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c21df91a9e4e3165686882b2eb559498224063dc633c3a63f958598108eca29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j24lb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:07:16Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-kl5s9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:08:01Z is after 2025-08-24T17:21:41Z" Dec 03 00:08:01 crc kubenswrapper[4953]: I1203 00:08:01.457719 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-xshk4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bd03a758-0ca9-47d8-86f7-18b2d610df43\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47gfh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47gfh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:07:31Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-xshk4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:08:01Z is after 2025-08-24T17:21:41Z" Dec 03 00:08:01 crc kubenswrapper[4953]: I1203 00:08:01.466567 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:01 crc kubenswrapper[4953]: I1203 00:08:01.466602 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:01 crc kubenswrapper[4953]: I1203 00:08:01.466614 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:01 crc kubenswrapper[4953]: I1203 00:08:01.466632 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:01 crc kubenswrapper[4953]: I1203 00:08:01.466645 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:01Z","lastTransitionTime":"2025-12-03T00:08:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:01 crc kubenswrapper[4953]: I1203 00:08:01.468932 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cf821840-5353-4620-8f73-99b56b371ab0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cae5e2994fcc24dc9b2ea73e79572615b3311edc42642b7f278fb84123989b23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://29e4b9d1b70a93fb133e0759ad148ae133766a66efeae3ff06a21dfae0e805ff\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dec43536772aa7c3f5a662fa7f3bd73b1cae2791ad454f85792287828d0ffeed\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://034e0bac6892d361700b8f33d6427b8a55e2bd3f89b14ea77a374c6be94d25e2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://761973393a93c59c23e69f41b8ac0d16abc35abd61ecbc6d51a42d94cd664233\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T00:07:09Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 00:07:03.713374 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 00:07:03.715052 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1528563048/tls.crt::/tmp/serving-cert-1528563048/tls.key\\\\\\\"\\\\nI1203 00:07:09.300052 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 00:07:09.304192 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 00:07:09.304255 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 00:07:09.304309 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 00:07:09.304323 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 00:07:09.313563 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 00:07:09.313597 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 00:07:09.313603 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 00:07:09.313609 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 00:07:09.313613 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 00:07:09.313618 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 00:07:09.313622 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 00:07:09.313709 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 00:07:09.327347 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T00:06:53Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c143d4232fb8bf7ef5999a664e431a00ab0db030f4d31a28d0bb21e00e38581\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:53Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72af6c27296eae2e275d455795c68f0d272869d63642b8353b17d466d0c901cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://72af6c27296eae2e275d455795c68f0d272869d63642b8353b17d466d0c901cf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:06:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:06:51Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:08:01Z is after 2025-08-24T17:21:41Z" Dec 03 00:08:01 crc kubenswrapper[4953]: I1203 00:08:01.483143 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:10Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:08:01Z is after 2025-08-24T17:21:41Z" Dec 03 00:08:01 crc kubenswrapper[4953]: I1203 00:08:01.498477 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce8b414e1fc22eb30decb6a6411b672e7f443055cb89ebb029fbf6ffc1d5dca7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:08:01Z is after 2025-08-24T17:21:41Z" Dec 03 00:08:01 crc kubenswrapper[4953]: I1203 00:08:01.511155 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-6dfbn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53ffb1e9-9a2e-4869-80d7-776f4aa7ef8b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ab4205c0d38ad6a0c32077eadb1fd5fd3a279d9c8b8a2ffbe693ec2c4018370c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ct26c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7df84405eda520ad5a22d35750edf07ef3c3977a756ad5739c54f91cba9da7b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ct26c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:07:16Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-6dfbn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:08:01Z is after 2025-08-24T17:21:41Z" Dec 03 00:08:01 crc kubenswrapper[4953]: I1203 00:08:01.536788 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-txffx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8816ebee-c735-4b8d-b1ec-65349707b347\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:16Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:16Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4b85224419a4075acb340e4b11095f2adb5c316e285afa4e19295c2181c0d24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0f6f645a58993f052a6f365bb808f62d99f7fea74886c5cc022b2d86112321a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a61b951a9cbab2c48c0f2df8351f97d7ff71bf3674f641fe5eac30155232b224\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd5f428de229749485267e81d34e31b708728b52474032e88460e54747f21e11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://00a2ea4a9bb870fce30e7fa845f787f184e7a19e42c95511655ea8c5f84a26d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc0a5d4cc9d614d8492626a1b114bdc52f510eb5965d842c02c083f08d4849a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dcb805f9ce285d3841e93d588fe7f3d42232382366d42b75bcfb129aefa91c1f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dcb805f9ce285d3841e93d588fe7f3d42232382366d42b75bcfb129aefa91c1f\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T00:07:44Z\\\",\\\"message\\\":\\\"mers/externalversions/factory.go:140\\\\nI1203 00:07:44.043157 6594 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1203 00:07:44.043232 6594 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 00:07:44.043450 6594 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 00:07:44.043811 6594 reflector.go:311] Stopping reflector *v1.NetworkAttachmentDefinition (0s) from github.com/k8snetworkplumbingwg/network-attachment-definition-client/pkg/client/informers/externalversions/factory.go:117\\\\nI1203 00:07:44.044028 6594 reflector.go:311] Stopping reflector *v1.AdminPolicyBasedExternalRoute (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140\\\\nI1203 00:07:44.044197 6594 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI1203 00:07:44.044257 6594 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T00:07:43Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-txffx_openshift-ovn-kubernetes(8816ebee-c735-4b8d-b1ec-65349707b347)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7deadc8e135ff1bc7c5a2f093e489bb5b4b6f166bc874c6acc2a841c5397f11e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af4c7764fca87144d01f3738ea010308744ccf50c54b7c471ccc213112de00d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af4c7764fca87144d01f3738ea010308744ccf50c54b7c471ccc213112de00d5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:07:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:07:16Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-txffx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:08:01Z is after 2025-08-24T17:21:41Z" Dec 03 00:08:01 crc kubenswrapper[4953]: I1203 00:08:01.548186 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5cmlc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7e11a1d3-ecee-4bf0-87bf-085fab97b73d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d42390b9c53efc63add8ddcebf03c976f5e5448eaa2c34e076b2dc90dbac60fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6x2sc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:07:18Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5cmlc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:08:01Z is after 2025-08-24T17:21:41Z" Dec 03 00:08:01 crc kubenswrapper[4953]: I1203 00:08:01.558168 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-chmnz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a12fe6d-d25a-4b08-90da-cc9e510980a7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://711ab839cb9c0a90554dadbd45d810fc78aac09739340482fef8de8743b77283\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c982g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f802d5628a8cc32618044b9383211d5a492c56060d427f953b7793a1064029e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c982g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:07:29Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-chmnz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:08:01Z is after 2025-08-24T17:21:41Z" Dec 03 00:08:01 crc kubenswrapper[4953]: I1203 00:08:01.567442 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7cf23d2d-21d4-493d-bdc7-d088aa876304\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4c95eb1f9b0c797376a1b557415b3f956890c971dc77c5257f238b4f32ff9b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ad2a6f5a3296c6a4f53f9da696e3e3f4d6c919f44b380c9b4114021b8ad74b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ad2a6f5a3296c6a4f53f9da696e3e3f4d6c919f44b380c9b4114021b8ad74b6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:06:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:06:51Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:08:01Z is after 2025-08-24T17:21:41Z" Dec 03 00:08:01 crc kubenswrapper[4953]: I1203 00:08:01.569897 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:01 crc kubenswrapper[4953]: I1203 00:08:01.569950 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:01 crc kubenswrapper[4953]: I1203 00:08:01.569961 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:01 crc kubenswrapper[4953]: I1203 00:08:01.569995 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:01 crc kubenswrapper[4953]: I1203 00:08:01.570007 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:01Z","lastTransitionTime":"2025-12-03T00:08:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:01 crc kubenswrapper[4953]: I1203 00:08:01.580155 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:08:01Z is after 2025-08-24T17:21:41Z" Dec 03 00:08:01 crc kubenswrapper[4953]: I1203 00:08:01.591750 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2ed9c7fa96c8951fbc1728fe7a91fefdd999077285621df9c1abedf9ae0c449\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:08:01Z is after 2025-08-24T17:21:41Z" Dec 03 00:08:01 crc kubenswrapper[4953]: I1203 00:08:01.673312 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:01 crc kubenswrapper[4953]: I1203 00:08:01.673372 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:01 crc kubenswrapper[4953]: I1203 00:08:01.673384 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:01 crc kubenswrapper[4953]: I1203 00:08:01.673402 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:01 crc kubenswrapper[4953]: I1203 00:08:01.673413 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:01Z","lastTransitionTime":"2025-12-03T00:08:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:01 crc kubenswrapper[4953]: I1203 00:08:01.776864 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:01 crc kubenswrapper[4953]: I1203 00:08:01.776916 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:01 crc kubenswrapper[4953]: I1203 00:08:01.776927 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:01 crc kubenswrapper[4953]: I1203 00:08:01.776944 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:01 crc kubenswrapper[4953]: I1203 00:08:01.776954 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:01Z","lastTransitionTime":"2025-12-03T00:08:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:01 crc kubenswrapper[4953]: I1203 00:08:01.881309 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:01 crc kubenswrapper[4953]: I1203 00:08:01.881606 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:01 crc kubenswrapper[4953]: I1203 00:08:01.881692 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:01 crc kubenswrapper[4953]: I1203 00:08:01.881790 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:01 crc kubenswrapper[4953]: I1203 00:08:01.881878 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:01Z","lastTransitionTime":"2025-12-03T00:08:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:01 crc kubenswrapper[4953]: I1203 00:08:01.984633 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:01 crc kubenswrapper[4953]: I1203 00:08:01.984689 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:01 crc kubenswrapper[4953]: I1203 00:08:01.984702 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:01 crc kubenswrapper[4953]: I1203 00:08:01.984724 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:01 crc kubenswrapper[4953]: I1203 00:08:01.984736 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:01Z","lastTransitionTime":"2025-12-03T00:08:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:02 crc kubenswrapper[4953]: I1203 00:08:02.088580 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:02 crc kubenswrapper[4953]: I1203 00:08:02.088624 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:02 crc kubenswrapper[4953]: I1203 00:08:02.088636 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:02 crc kubenswrapper[4953]: I1203 00:08:02.088691 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:02 crc kubenswrapper[4953]: I1203 00:08:02.088706 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:02Z","lastTransitionTime":"2025-12-03T00:08:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:02 crc kubenswrapper[4953]: I1203 00:08:02.191250 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:02 crc kubenswrapper[4953]: I1203 00:08:02.191293 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:02 crc kubenswrapper[4953]: I1203 00:08:02.191302 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:02 crc kubenswrapper[4953]: I1203 00:08:02.191319 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:02 crc kubenswrapper[4953]: I1203 00:08:02.191330 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:02Z","lastTransitionTime":"2025-12-03T00:08:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:02 crc kubenswrapper[4953]: I1203 00:08:02.293553 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:02 crc kubenswrapper[4953]: I1203 00:08:02.293633 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:02 crc kubenswrapper[4953]: I1203 00:08:02.293656 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:02 crc kubenswrapper[4953]: I1203 00:08:02.293686 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:02 crc kubenswrapper[4953]: I1203 00:08:02.293711 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:02Z","lastTransitionTime":"2025-12-03T00:08:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:02 crc kubenswrapper[4953]: I1203 00:08:02.347495 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 00:08:02 crc kubenswrapper[4953]: I1203 00:08:02.347520 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 00:08:02 crc kubenswrapper[4953]: E1203 00:08:02.347685 4953 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 00:08:02 crc kubenswrapper[4953]: I1203 00:08:02.347610 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 00:08:02 crc kubenswrapper[4953]: E1203 00:08:02.347777 4953 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 00:08:02 crc kubenswrapper[4953]: E1203 00:08:02.347949 4953 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 00:08:02 crc kubenswrapper[4953]: I1203 00:08:02.397481 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:02 crc kubenswrapper[4953]: I1203 00:08:02.397539 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:02 crc kubenswrapper[4953]: I1203 00:08:02.397551 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:02 crc kubenswrapper[4953]: I1203 00:08:02.397576 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:02 crc kubenswrapper[4953]: I1203 00:08:02.397589 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:02Z","lastTransitionTime":"2025-12-03T00:08:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:02 crc kubenswrapper[4953]: I1203 00:08:02.500630 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:02 crc kubenswrapper[4953]: I1203 00:08:02.500695 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:02 crc kubenswrapper[4953]: I1203 00:08:02.500709 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:02 crc kubenswrapper[4953]: I1203 00:08:02.500730 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:02 crc kubenswrapper[4953]: I1203 00:08:02.500744 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:02Z","lastTransitionTime":"2025-12-03T00:08:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:02 crc kubenswrapper[4953]: I1203 00:08:02.603902 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:02 crc kubenswrapper[4953]: I1203 00:08:02.603953 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:02 crc kubenswrapper[4953]: I1203 00:08:02.603965 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:02 crc kubenswrapper[4953]: I1203 00:08:02.604012 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:02 crc kubenswrapper[4953]: I1203 00:08:02.604026 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:02Z","lastTransitionTime":"2025-12-03T00:08:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:02 crc kubenswrapper[4953]: I1203 00:08:02.707483 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:02 crc kubenswrapper[4953]: I1203 00:08:02.707537 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:02 crc kubenswrapper[4953]: I1203 00:08:02.707549 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:02 crc kubenswrapper[4953]: I1203 00:08:02.707571 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:02 crc kubenswrapper[4953]: I1203 00:08:02.707582 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:02Z","lastTransitionTime":"2025-12-03T00:08:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:02 crc kubenswrapper[4953]: I1203 00:08:02.811103 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:02 crc kubenswrapper[4953]: I1203 00:08:02.811139 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:02 crc kubenswrapper[4953]: I1203 00:08:02.811148 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:02 crc kubenswrapper[4953]: I1203 00:08:02.811171 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:02 crc kubenswrapper[4953]: I1203 00:08:02.811183 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:02Z","lastTransitionTime":"2025-12-03T00:08:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:02 crc kubenswrapper[4953]: I1203 00:08:02.830029 4953 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-675qk_43a4bad8-71f4-441a-8302-aeb526a986af/kube-multus/0.log" Dec 03 00:08:02 crc kubenswrapper[4953]: I1203 00:08:02.830108 4953 generic.go:334] "Generic (PLEG): container finished" podID="43a4bad8-71f4-441a-8302-aeb526a986af" containerID="5b5794fef6477aab5deed8f642e8cc50b37eb1fd0cc6f19c003f34ad31c1e3c0" exitCode=1 Dec 03 00:08:02 crc kubenswrapper[4953]: I1203 00:08:02.830156 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-675qk" event={"ID":"43a4bad8-71f4-441a-8302-aeb526a986af","Type":"ContainerDied","Data":"5b5794fef6477aab5deed8f642e8cc50b37eb1fd0cc6f19c003f34ad31c1e3c0"} Dec 03 00:08:02 crc kubenswrapper[4953]: I1203 00:08:02.830789 4953 scope.go:117] "RemoveContainer" containerID="5b5794fef6477aab5deed8f642e8cc50b37eb1fd0cc6f19c003f34ad31c1e3c0" Dec 03 00:08:02 crc kubenswrapper[4953]: I1203 00:08:02.844768 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:10Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:08:02Z is after 2025-08-24T17:21:41Z" Dec 03 00:08:02 crc kubenswrapper[4953]: I1203 00:08:02.861163 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-675qk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"43a4bad8-71f4-441a-8302-aeb526a986af\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:08:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:08:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5b5794fef6477aab5deed8f642e8cc50b37eb1fd0cc6f19c003f34ad31c1e3c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5b5794fef6477aab5deed8f642e8cc50b37eb1fd0cc6f19c003f34ad31c1e3c0\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T00:08:02Z\\\",\\\"message\\\":\\\"2025-12-03T00:07:17+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_fa87bd16-684d-458f-b6b8-b6ada86b0d46\\\\n2025-12-03T00:07:17+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_fa87bd16-684d-458f-b6b8-b6ada86b0d46 to /host/opt/cni/bin/\\\\n2025-12-03T00:07:17Z [verbose] multus-daemon started\\\\n2025-12-03T00:07:17Z [verbose] Readiness Indicator file check\\\\n2025-12-03T00:08:02Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T00:07:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gdrw4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:07:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-675qk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:08:02Z is after 2025-08-24T17:21:41Z" Dec 03 00:08:02 crc kubenswrapper[4953]: I1203 00:08:02.875509 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-kl5s9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d85e5c97-16ad-4bd7-9e73-b7f12678be97\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c21df91a9e4e3165686882b2eb559498224063dc633c3a63f958598108eca29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j24lb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:07:16Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-kl5s9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:08:02Z is after 2025-08-24T17:21:41Z" Dec 03 00:08:02 crc kubenswrapper[4953]: I1203 00:08:02.891070 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f94f8c51-c665-4ca3-bf17-7a04ea5c7ade\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72c5cd7dc1c3a8adfedab6ba43a59d80b3efe65f2bb56be95e76cc86a409dcf3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2810ffc1b141fa7da811406053db9d72c708edf5f2bb1dd75e4d52ab35fd24c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c67b4277b20a9f1287f50ef82d0785c9d02543a8e7ed26e0413eae332b866db7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://258d78e1ad9d3dcda6976909ae3fe71aecead5b995fc66b3fd6e2722d80b5ab8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:06:51Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:08:02Z is after 2025-08-24T17:21:41Z" Dec 03 00:08:02 crc kubenswrapper[4953]: I1203 00:08:02.908958 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5652d0f3-3fd7-4af6-8941-9119e1177641\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2030aef5b7170991268fa5dc7e7ed8ce1fbe3a2765165a10c25251426575330f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ab6052f1abf17225155ce1bd86afa742e6e78f507b7deedd7161cd8fba9120b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://650b6c79cda2e1768ba1d8763db799e210ce2748893512068d25e60f821937ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d1d60162f7f6885d4f51b4ab44163cfae6d2d04355b1dd710e26b022450005df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d1d60162f7f6885d4f51b4ab44163cfae6d2d04355b1dd710e26b022450005df\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:06:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:06:51Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:08:02Z is after 2025-08-24T17:21:41Z" Dec 03 00:08:02 crc kubenswrapper[4953]: I1203 00:08:02.915699 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:02 crc kubenswrapper[4953]: I1203 00:08:02.915725 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:02 crc kubenswrapper[4953]: I1203 00:08:02.915738 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:02 crc kubenswrapper[4953]: I1203 00:08:02.915760 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:02 crc kubenswrapper[4953]: I1203 00:08:02.915778 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:02Z","lastTransitionTime":"2025-12-03T00:08:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:02 crc kubenswrapper[4953]: I1203 00:08:02.963373 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09657a9f448a4c92a759c73a0a892076d7eb46393e6fd565e5f73f34c71aac23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5af5eac4a58870f63497f99b503cfa5fb922278095e0332ec50376925ecf2109\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:08:02Z is after 2025-08-24T17:21:41Z" Dec 03 00:08:02 crc kubenswrapper[4953]: I1203 00:08:02.989478 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-wj6rz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"96aa67a0-15dd-420d-a1d5-ed239d7df1fb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01815cfd3f58c3f3f5f2a894c1790e0b69c5060c48dc0b1deff050b6ad7af916\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kf62m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b12bfb5bdf9fd4f2c50a335fc82df6ee0440eaf8741897da03aa2dae368cc6b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b12bfb5bdf9fd4f2c50a335fc82df6ee0440eaf8741897da03aa2dae368cc6b4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:07:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kf62m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83aaa5fa28110bc0ef05aff75cd1442bc9bbdd33b822ae2ef22d4e4ab83238bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://83aaa5fa28110bc0ef05aff75cd1442bc9bbdd33b822ae2ef22d4e4ab83238bc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:07:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:07:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kf62m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ab6213a33a19fedebce6d74a0ebbb3b3a2de5687504e21cc24093cc42ade635\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ab6213a33a19fedebce6d74a0ebbb3b3a2de5687504e21cc24093cc42ade635\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:07:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:07:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kf62m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://de480488843afa6b8dede1e1b42b5181dca1993e2f39dc1b461a18a57e3668e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://de480488843afa6b8dede1e1b42b5181dca1993e2f39dc1b461a18a57e3668e1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:07:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kf62m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://97be4ba692bc61078d4b0ba39fb0c12772548c6c9251ff3356ea702f151223cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://97be4ba692bc61078d4b0ba39fb0c12772548c6c9251ff3356ea702f151223cd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:07:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kf62m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1a1f009fcbd8af9217db1dc97f662719f8c1db232df0e2804e44c881abb1df2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b1a1f009fcbd8af9217db1dc97f662719f8c1db232df0e2804e44c881abb1df2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:07:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:07:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kf62m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:07:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-wj6rz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:08:02Z is after 2025-08-24T17:21:41Z" Dec 03 00:08:03 crc kubenswrapper[4953]: I1203 00:08:03.010317 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-txffx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8816ebee-c735-4b8d-b1ec-65349707b347\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:16Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:16Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4b85224419a4075acb340e4b11095f2adb5c316e285afa4e19295c2181c0d24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0f6f645a58993f052a6f365bb808f62d99f7fea74886c5cc022b2d86112321a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a61b951a9cbab2c48c0f2df8351f97d7ff71bf3674f641fe5eac30155232b224\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd5f428de229749485267e81d34e31b708728b52474032e88460e54747f21e11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://00a2ea4a9bb870fce30e7fa845f787f184e7a19e42c95511655ea8c5f84a26d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc0a5d4cc9d614d8492626a1b114bdc52f510eb5965d842c02c083f08d4849a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dcb805f9ce285d3841e93d588fe7f3d42232382366d42b75bcfb129aefa91c1f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dcb805f9ce285d3841e93d588fe7f3d42232382366d42b75bcfb129aefa91c1f\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T00:07:44Z\\\",\\\"message\\\":\\\"mers/externalversions/factory.go:140\\\\nI1203 00:07:44.043157 6594 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1203 00:07:44.043232 6594 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 00:07:44.043450 6594 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 00:07:44.043811 6594 reflector.go:311] Stopping reflector *v1.NetworkAttachmentDefinition (0s) from github.com/k8snetworkplumbingwg/network-attachment-definition-client/pkg/client/informers/externalversions/factory.go:117\\\\nI1203 00:07:44.044028 6594 reflector.go:311] Stopping reflector *v1.AdminPolicyBasedExternalRoute (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140\\\\nI1203 00:07:44.044197 6594 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI1203 00:07:44.044257 6594 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T00:07:43Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-txffx_openshift-ovn-kubernetes(8816ebee-c735-4b8d-b1ec-65349707b347)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7deadc8e135ff1bc7c5a2f093e489bb5b4b6f166bc874c6acc2a841c5397f11e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af4c7764fca87144d01f3738ea010308744ccf50c54b7c471ccc213112de00d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af4c7764fca87144d01f3738ea010308744ccf50c54b7c471ccc213112de00d5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:07:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:07:16Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-txffx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:08:03Z is after 2025-08-24T17:21:41Z" Dec 03 00:08:03 crc kubenswrapper[4953]: I1203 00:08:03.018396 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:03 crc kubenswrapper[4953]: I1203 00:08:03.018427 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:03 crc kubenswrapper[4953]: I1203 00:08:03.018435 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:03 crc kubenswrapper[4953]: I1203 00:08:03.018450 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:03 crc kubenswrapper[4953]: I1203 00:08:03.018461 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:03Z","lastTransitionTime":"2025-12-03T00:08:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:03 crc kubenswrapper[4953]: I1203 00:08:03.025064 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5cmlc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7e11a1d3-ecee-4bf0-87bf-085fab97b73d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d42390b9c53efc63add8ddcebf03c976f5e5448eaa2c34e076b2dc90dbac60fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6x2sc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:07:18Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5cmlc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:08:03Z is after 2025-08-24T17:21:41Z" Dec 03 00:08:03 crc kubenswrapper[4953]: I1203 00:08:03.038785 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-chmnz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a12fe6d-d25a-4b08-90da-cc9e510980a7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://711ab839cb9c0a90554dadbd45d810fc78aac09739340482fef8de8743b77283\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c982g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f802d5628a8cc32618044b9383211d5a492c56060d427f953b7793a1064029e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c982g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:07:29Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-chmnz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:08:03Z is after 2025-08-24T17:21:41Z" Dec 03 00:08:03 crc kubenswrapper[4953]: I1203 00:08:03.043888 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/bd03a758-0ca9-47d8-86f7-18b2d610df43-metrics-certs\") pod \"network-metrics-daemon-xshk4\" (UID: \"bd03a758-0ca9-47d8-86f7-18b2d610df43\") " pod="openshift-multus/network-metrics-daemon-xshk4" Dec 03 00:08:03 crc kubenswrapper[4953]: E1203 00:08:03.044091 4953 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 03 00:08:03 crc kubenswrapper[4953]: E1203 00:08:03.044153 4953 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/bd03a758-0ca9-47d8-86f7-18b2d610df43-metrics-certs podName:bd03a758-0ca9-47d8-86f7-18b2d610df43 nodeName:}" failed. No retries permitted until 2025-12-03 00:08:35.044131461 +0000 UTC m=+104.028412264 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/bd03a758-0ca9-47d8-86f7-18b2d610df43-metrics-certs") pod "network-metrics-daemon-xshk4" (UID: "bd03a758-0ca9-47d8-86f7-18b2d610df43") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 03 00:08:03 crc kubenswrapper[4953]: I1203 00:08:03.050105 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-xshk4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bd03a758-0ca9-47d8-86f7-18b2d610df43\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47gfh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47gfh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:07:31Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-xshk4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:08:03Z is after 2025-08-24T17:21:41Z" Dec 03 00:08:03 crc kubenswrapper[4953]: I1203 00:08:03.068248 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cf821840-5353-4620-8f73-99b56b371ab0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cae5e2994fcc24dc9b2ea73e79572615b3311edc42642b7f278fb84123989b23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://29e4b9d1b70a93fb133e0759ad148ae133766a66efeae3ff06a21dfae0e805ff\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dec43536772aa7c3f5a662fa7f3bd73b1cae2791ad454f85792287828d0ffeed\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://034e0bac6892d361700b8f33d6427b8a55e2bd3f89b14ea77a374c6be94d25e2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://761973393a93c59c23e69f41b8ac0d16abc35abd61ecbc6d51a42d94cd664233\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T00:07:09Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 00:07:03.713374 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 00:07:03.715052 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1528563048/tls.crt::/tmp/serving-cert-1528563048/tls.key\\\\\\\"\\\\nI1203 00:07:09.300052 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 00:07:09.304192 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 00:07:09.304255 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 00:07:09.304309 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 00:07:09.304323 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 00:07:09.313563 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 00:07:09.313597 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 00:07:09.313603 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 00:07:09.313609 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 00:07:09.313613 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 00:07:09.313618 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 00:07:09.313622 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 00:07:09.313709 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 00:07:09.327347 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T00:06:53Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c143d4232fb8bf7ef5999a664e431a00ab0db030f4d31a28d0bb21e00e38581\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:53Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72af6c27296eae2e275d455795c68f0d272869d63642b8353b17d466d0c901cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://72af6c27296eae2e275d455795c68f0d272869d63642b8353b17d466d0c901cf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:06:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:06:51Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:08:03Z is after 2025-08-24T17:21:41Z" Dec 03 00:08:03 crc kubenswrapper[4953]: I1203 00:08:03.084754 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:10Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:08:03Z is after 2025-08-24T17:21:41Z" Dec 03 00:08:03 crc kubenswrapper[4953]: I1203 00:08:03.102943 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce8b414e1fc22eb30decb6a6411b672e7f443055cb89ebb029fbf6ffc1d5dca7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:08:03Z is after 2025-08-24T17:21:41Z" Dec 03 00:08:03 crc kubenswrapper[4953]: I1203 00:08:03.115279 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-6dfbn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53ffb1e9-9a2e-4869-80d7-776f4aa7ef8b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ab4205c0d38ad6a0c32077eadb1fd5fd3a279d9c8b8a2ffbe693ec2c4018370c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ct26c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7df84405eda520ad5a22d35750edf07ef3c3977a756ad5739c54f91cba9da7b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ct26c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:07:16Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-6dfbn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:08:03Z is after 2025-08-24T17:21:41Z" Dec 03 00:08:03 crc kubenswrapper[4953]: I1203 00:08:03.120720 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:03 crc kubenswrapper[4953]: I1203 00:08:03.120750 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:03 crc kubenswrapper[4953]: I1203 00:08:03.120759 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:03 crc kubenswrapper[4953]: I1203 00:08:03.120773 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:03 crc kubenswrapper[4953]: I1203 00:08:03.120784 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:03Z","lastTransitionTime":"2025-12-03T00:08:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:03 crc kubenswrapper[4953]: I1203 00:08:03.126253 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7cf23d2d-21d4-493d-bdc7-d088aa876304\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4c95eb1f9b0c797376a1b557415b3f956890c971dc77c5257f238b4f32ff9b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ad2a6f5a3296c6a4f53f9da696e3e3f4d6c919f44b380c9b4114021b8ad74b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ad2a6f5a3296c6a4f53f9da696e3e3f4d6c919f44b380c9b4114021b8ad74b6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:06:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:06:51Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:08:03Z is after 2025-08-24T17:21:41Z" Dec 03 00:08:03 crc kubenswrapper[4953]: I1203 00:08:03.137706 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:08:03Z is after 2025-08-24T17:21:41Z" Dec 03 00:08:03 crc kubenswrapper[4953]: I1203 00:08:03.149749 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2ed9c7fa96c8951fbc1728fe7a91fefdd999077285621df9c1abedf9ae0c449\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:08:03Z is after 2025-08-24T17:21:41Z" Dec 03 00:08:03 crc kubenswrapper[4953]: I1203 00:08:03.222921 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:03 crc kubenswrapper[4953]: I1203 00:08:03.222969 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:03 crc kubenswrapper[4953]: I1203 00:08:03.223220 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:03 crc kubenswrapper[4953]: I1203 00:08:03.223241 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:03 crc kubenswrapper[4953]: I1203 00:08:03.223494 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:03Z","lastTransitionTime":"2025-12-03T00:08:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:03 crc kubenswrapper[4953]: I1203 00:08:03.326710 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:03 crc kubenswrapper[4953]: I1203 00:08:03.326756 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:03 crc kubenswrapper[4953]: I1203 00:08:03.326765 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:03 crc kubenswrapper[4953]: I1203 00:08:03.326780 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:03 crc kubenswrapper[4953]: I1203 00:08:03.326789 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:03Z","lastTransitionTime":"2025-12-03T00:08:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:03 crc kubenswrapper[4953]: I1203 00:08:03.347676 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xshk4" Dec 03 00:08:03 crc kubenswrapper[4953]: E1203 00:08:03.347836 4953 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xshk4" podUID="bd03a758-0ca9-47d8-86f7-18b2d610df43" Dec 03 00:08:03 crc kubenswrapper[4953]: I1203 00:08:03.429851 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:03 crc kubenswrapper[4953]: I1203 00:08:03.429929 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:03 crc kubenswrapper[4953]: I1203 00:08:03.429940 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:03 crc kubenswrapper[4953]: I1203 00:08:03.429961 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:03 crc kubenswrapper[4953]: I1203 00:08:03.429998 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:03Z","lastTransitionTime":"2025-12-03T00:08:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:03 crc kubenswrapper[4953]: I1203 00:08:03.532046 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:03 crc kubenswrapper[4953]: I1203 00:08:03.532413 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:03 crc kubenswrapper[4953]: I1203 00:08:03.532557 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:03 crc kubenswrapper[4953]: I1203 00:08:03.532717 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:03 crc kubenswrapper[4953]: I1203 00:08:03.532890 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:03Z","lastTransitionTime":"2025-12-03T00:08:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:03 crc kubenswrapper[4953]: I1203 00:08:03.635224 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:03 crc kubenswrapper[4953]: I1203 00:08:03.635273 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:03 crc kubenswrapper[4953]: I1203 00:08:03.635284 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:03 crc kubenswrapper[4953]: I1203 00:08:03.635299 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:03 crc kubenswrapper[4953]: I1203 00:08:03.635310 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:03Z","lastTransitionTime":"2025-12-03T00:08:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:03 crc kubenswrapper[4953]: I1203 00:08:03.738096 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:03 crc kubenswrapper[4953]: I1203 00:08:03.738251 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:03 crc kubenswrapper[4953]: I1203 00:08:03.738360 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:03 crc kubenswrapper[4953]: I1203 00:08:03.738454 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:03 crc kubenswrapper[4953]: I1203 00:08:03.738542 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:03Z","lastTransitionTime":"2025-12-03T00:08:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:03 crc kubenswrapper[4953]: I1203 00:08:03.834911 4953 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-675qk_43a4bad8-71f4-441a-8302-aeb526a986af/kube-multus/0.log" Dec 03 00:08:03 crc kubenswrapper[4953]: I1203 00:08:03.835052 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-675qk" event={"ID":"43a4bad8-71f4-441a-8302-aeb526a986af","Type":"ContainerStarted","Data":"e701c2ad05014683415ae2fb03b96f3acb38062ba7c5bfc4d0a9e34431695e36"} Dec 03 00:08:03 crc kubenswrapper[4953]: I1203 00:08:03.840286 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:03 crc kubenswrapper[4953]: I1203 00:08:03.840347 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:03 crc kubenswrapper[4953]: I1203 00:08:03.840371 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:03 crc kubenswrapper[4953]: I1203 00:08:03.840400 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:03 crc kubenswrapper[4953]: I1203 00:08:03.840425 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:03Z","lastTransitionTime":"2025-12-03T00:08:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:03 crc kubenswrapper[4953]: I1203 00:08:03.857195 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce8b414e1fc22eb30decb6a6411b672e7f443055cb89ebb029fbf6ffc1d5dca7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:08:03Z is after 2025-08-24T17:21:41Z" Dec 03 00:08:03 crc kubenswrapper[4953]: I1203 00:08:03.877033 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-6dfbn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53ffb1e9-9a2e-4869-80d7-776f4aa7ef8b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ab4205c0d38ad6a0c32077eadb1fd5fd3a279d9c8b8a2ffbe693ec2c4018370c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ct26c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7df84405eda520ad5a22d35750edf07ef3c3977a756ad5739c54f91cba9da7b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ct26c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:07:16Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-6dfbn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:08:03Z is after 2025-08-24T17:21:41Z" Dec 03 00:08:03 crc kubenswrapper[4953]: I1203 00:08:03.910731 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-txffx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8816ebee-c735-4b8d-b1ec-65349707b347\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:16Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:16Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4b85224419a4075acb340e4b11095f2adb5c316e285afa4e19295c2181c0d24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0f6f645a58993f052a6f365bb808f62d99f7fea74886c5cc022b2d86112321a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a61b951a9cbab2c48c0f2df8351f97d7ff71bf3674f641fe5eac30155232b224\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd5f428de229749485267e81d34e31b708728b52474032e88460e54747f21e11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://00a2ea4a9bb870fce30e7fa845f787f184e7a19e42c95511655ea8c5f84a26d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc0a5d4cc9d614d8492626a1b114bdc52f510eb5965d842c02c083f08d4849a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dcb805f9ce285d3841e93d588fe7f3d42232382366d42b75bcfb129aefa91c1f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dcb805f9ce285d3841e93d588fe7f3d42232382366d42b75bcfb129aefa91c1f\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T00:07:44Z\\\",\\\"message\\\":\\\"mers/externalversions/factory.go:140\\\\nI1203 00:07:44.043157 6594 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1203 00:07:44.043232 6594 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 00:07:44.043450 6594 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 00:07:44.043811 6594 reflector.go:311] Stopping reflector *v1.NetworkAttachmentDefinition (0s) from github.com/k8snetworkplumbingwg/network-attachment-definition-client/pkg/client/informers/externalversions/factory.go:117\\\\nI1203 00:07:44.044028 6594 reflector.go:311] Stopping reflector *v1.AdminPolicyBasedExternalRoute (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140\\\\nI1203 00:07:44.044197 6594 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI1203 00:07:44.044257 6594 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T00:07:43Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-txffx_openshift-ovn-kubernetes(8816ebee-c735-4b8d-b1ec-65349707b347)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7deadc8e135ff1bc7c5a2f093e489bb5b4b6f166bc874c6acc2a841c5397f11e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af4c7764fca87144d01f3738ea010308744ccf50c54b7c471ccc213112de00d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af4c7764fca87144d01f3738ea010308744ccf50c54b7c471ccc213112de00d5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:07:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:07:16Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-txffx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:08:03Z is after 2025-08-24T17:21:41Z" Dec 03 00:08:03 crc kubenswrapper[4953]: I1203 00:08:03.929785 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5cmlc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7e11a1d3-ecee-4bf0-87bf-085fab97b73d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d42390b9c53efc63add8ddcebf03c976f5e5448eaa2c34e076b2dc90dbac60fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6x2sc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:07:18Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5cmlc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:08:03Z is after 2025-08-24T17:21:41Z" Dec 03 00:08:03 crc kubenswrapper[4953]: I1203 00:08:03.945232 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:03 crc kubenswrapper[4953]: I1203 00:08:03.945265 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:03 crc kubenswrapper[4953]: I1203 00:08:03.945273 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:03 crc kubenswrapper[4953]: I1203 00:08:03.945289 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:03 crc kubenswrapper[4953]: I1203 00:08:03.945298 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:03Z","lastTransitionTime":"2025-12-03T00:08:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:03 crc kubenswrapper[4953]: I1203 00:08:03.952066 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-chmnz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a12fe6d-d25a-4b08-90da-cc9e510980a7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://711ab839cb9c0a90554dadbd45d810fc78aac09739340482fef8de8743b77283\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c982g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f802d5628a8cc32618044b9383211d5a492c56060d427f953b7793a1064029e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c982g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:07:29Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-chmnz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:08:03Z is after 2025-08-24T17:21:41Z" Dec 03 00:08:03 crc kubenswrapper[4953]: I1203 00:08:03.968953 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-xshk4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bd03a758-0ca9-47d8-86f7-18b2d610df43\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47gfh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47gfh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:07:31Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-xshk4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:08:03Z is after 2025-08-24T17:21:41Z" Dec 03 00:08:03 crc kubenswrapper[4953]: I1203 00:08:03.989085 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cf821840-5353-4620-8f73-99b56b371ab0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cae5e2994fcc24dc9b2ea73e79572615b3311edc42642b7f278fb84123989b23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://29e4b9d1b70a93fb133e0759ad148ae133766a66efeae3ff06a21dfae0e805ff\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dec43536772aa7c3f5a662fa7f3bd73b1cae2791ad454f85792287828d0ffeed\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://034e0bac6892d361700b8f33d6427b8a55e2bd3f89b14ea77a374c6be94d25e2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://761973393a93c59c23e69f41b8ac0d16abc35abd61ecbc6d51a42d94cd664233\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T00:07:09Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 00:07:03.713374 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 00:07:03.715052 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1528563048/tls.crt::/tmp/serving-cert-1528563048/tls.key\\\\\\\"\\\\nI1203 00:07:09.300052 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 00:07:09.304192 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 00:07:09.304255 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 00:07:09.304309 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 00:07:09.304323 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 00:07:09.313563 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 00:07:09.313597 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 00:07:09.313603 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 00:07:09.313609 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 00:07:09.313613 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 00:07:09.313618 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 00:07:09.313622 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 00:07:09.313709 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 00:07:09.327347 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T00:06:53Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c143d4232fb8bf7ef5999a664e431a00ab0db030f4d31a28d0bb21e00e38581\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:53Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72af6c27296eae2e275d455795c68f0d272869d63642b8353b17d466d0c901cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://72af6c27296eae2e275d455795c68f0d272869d63642b8353b17d466d0c901cf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:06:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:06:51Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:08:03Z is after 2025-08-24T17:21:41Z" Dec 03 00:08:04 crc kubenswrapper[4953]: I1203 00:08:04.007532 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:10Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:08:04Z is after 2025-08-24T17:21:41Z" Dec 03 00:08:04 crc kubenswrapper[4953]: I1203 00:08:04.023704 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:08:04Z is after 2025-08-24T17:21:41Z" Dec 03 00:08:04 crc kubenswrapper[4953]: I1203 00:08:04.050146 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:04 crc kubenswrapper[4953]: I1203 00:08:04.050207 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:04 crc kubenswrapper[4953]: I1203 00:08:04.050224 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:04 crc kubenswrapper[4953]: I1203 00:08:04.050251 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:04 crc kubenswrapper[4953]: I1203 00:08:04.050270 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:04Z","lastTransitionTime":"2025-12-03T00:08:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:04 crc kubenswrapper[4953]: I1203 00:08:04.076326 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2ed9c7fa96c8951fbc1728fe7a91fefdd999077285621df9c1abedf9ae0c449\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:08:04Z is after 2025-08-24T17:21:41Z" Dec 03 00:08:04 crc kubenswrapper[4953]: I1203 00:08:04.088452 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7cf23d2d-21d4-493d-bdc7-d088aa876304\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4c95eb1f9b0c797376a1b557415b3f956890c971dc77c5257f238b4f32ff9b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ad2a6f5a3296c6a4f53f9da696e3e3f4d6c919f44b380c9b4114021b8ad74b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ad2a6f5a3296c6a4f53f9da696e3e3f4d6c919f44b380c9b4114021b8ad74b6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:06:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:06:51Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:08:04Z is after 2025-08-24T17:21:41Z" Dec 03 00:08:04 crc kubenswrapper[4953]: I1203 00:08:04.102039 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:10Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:08:04Z is after 2025-08-24T17:21:41Z" Dec 03 00:08:04 crc kubenswrapper[4953]: I1203 00:08:04.114063 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-675qk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"43a4bad8-71f4-441a-8302-aeb526a986af\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:08:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:08:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e701c2ad05014683415ae2fb03b96f3acb38062ba7c5bfc4d0a9e34431695e36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5b5794fef6477aab5deed8f642e8cc50b37eb1fd0cc6f19c003f34ad31c1e3c0\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T00:08:02Z\\\",\\\"message\\\":\\\"2025-12-03T00:07:17+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_fa87bd16-684d-458f-b6b8-b6ada86b0d46\\\\n2025-12-03T00:07:17+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_fa87bd16-684d-458f-b6b8-b6ada86b0d46 to /host/opt/cni/bin/\\\\n2025-12-03T00:07:17Z [verbose] multus-daemon started\\\\n2025-12-03T00:07:17Z [verbose] Readiness Indicator file check\\\\n2025-12-03T00:08:02Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T00:07:16Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:08:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gdrw4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:07:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-675qk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:08:04Z is after 2025-08-24T17:21:41Z" Dec 03 00:08:04 crc kubenswrapper[4953]: I1203 00:08:04.124252 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:04 crc kubenswrapper[4953]: I1203 00:08:04.124304 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:04 crc kubenswrapper[4953]: I1203 00:08:04.124321 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:04 crc kubenswrapper[4953]: I1203 00:08:04.124344 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:04 crc kubenswrapper[4953]: I1203 00:08:04.124360 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:04Z","lastTransitionTime":"2025-12-03T00:08:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:04 crc kubenswrapper[4953]: I1203 00:08:04.127279 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09657a9f448a4c92a759c73a0a892076d7eb46393e6fd565e5f73f34c71aac23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5af5eac4a58870f63497f99b503cfa5fb922278095e0332ec50376925ecf2109\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:08:04Z is after 2025-08-24T17:21:41Z" Dec 03 00:08:04 crc kubenswrapper[4953]: E1203 00:08:04.144586 4953 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:08:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:08:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:08:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:08:04Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:08:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:08:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:08:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:08:04Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"eb4ae5a6-0f6f-49ef-a027-1cdee9d3c536\\\",\\\"systemUUID\\\":\\\"4c1f3d44-3d3d-425d-993f-471223206a89\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:08:04Z is after 2025-08-24T17:21:41Z" Dec 03 00:08:04 crc kubenswrapper[4953]: I1203 00:08:04.146706 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-wj6rz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"96aa67a0-15dd-420d-a1d5-ed239d7df1fb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01815cfd3f58c3f3f5f2a894c1790e0b69c5060c48dc0b1deff050b6ad7af916\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kf62m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b12bfb5bdf9fd4f2c50a335fc82df6ee0440eaf8741897da03aa2dae368cc6b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b12bfb5bdf9fd4f2c50a335fc82df6ee0440eaf8741897da03aa2dae368cc6b4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:07:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kf62m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83aaa5fa28110bc0ef05aff75cd1442bc9bbdd33b822ae2ef22d4e4ab83238bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://83aaa5fa28110bc0ef05aff75cd1442bc9bbdd33b822ae2ef22d4e4ab83238bc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:07:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:07:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kf62m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ab6213a33a19fedebce6d74a0ebbb3b3a2de5687504e21cc24093cc42ade635\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ab6213a33a19fedebce6d74a0ebbb3b3a2de5687504e21cc24093cc42ade635\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:07:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:07:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kf62m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://de480488843afa6b8dede1e1b42b5181dca1993e2f39dc1b461a18a57e3668e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://de480488843afa6b8dede1e1b42b5181dca1993e2f39dc1b461a18a57e3668e1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:07:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kf62m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://97be4ba692bc61078d4b0ba39fb0c12772548c6c9251ff3356ea702f151223cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://97be4ba692bc61078d4b0ba39fb0c12772548c6c9251ff3356ea702f151223cd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:07:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kf62m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1a1f009fcbd8af9217db1dc97f662719f8c1db232df0e2804e44c881abb1df2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b1a1f009fcbd8af9217db1dc97f662719f8c1db232df0e2804e44c881abb1df2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:07:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:07:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kf62m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:07:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-wj6rz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:08:04Z is after 2025-08-24T17:21:41Z" Dec 03 00:08:04 crc kubenswrapper[4953]: I1203 00:08:04.149556 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:04 crc kubenswrapper[4953]: I1203 00:08:04.149583 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:04 crc kubenswrapper[4953]: I1203 00:08:04.149595 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:04 crc kubenswrapper[4953]: I1203 00:08:04.149612 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:04 crc kubenswrapper[4953]: I1203 00:08:04.149625 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:04Z","lastTransitionTime":"2025-12-03T00:08:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:04 crc kubenswrapper[4953]: I1203 00:08:04.161074 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-kl5s9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d85e5c97-16ad-4bd7-9e73-b7f12678be97\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c21df91a9e4e3165686882b2eb559498224063dc633c3a63f958598108eca29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j24lb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:07:16Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-kl5s9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:08:04Z is after 2025-08-24T17:21:41Z" Dec 03 00:08:04 crc kubenswrapper[4953]: E1203 00:08:04.168526 4953 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:08:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:08:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:08:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:08:04Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:08:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:08:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:08:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:08:04Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"eb4ae5a6-0f6f-49ef-a027-1cdee9d3c536\\\",\\\"systemUUID\\\":\\\"4c1f3d44-3d3d-425d-993f-471223206a89\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:08:04Z is after 2025-08-24T17:21:41Z" Dec 03 00:08:04 crc kubenswrapper[4953]: I1203 00:08:04.172902 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:04 crc kubenswrapper[4953]: I1203 00:08:04.172985 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:04 crc kubenswrapper[4953]: I1203 00:08:04.173003 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:04 crc kubenswrapper[4953]: I1203 00:08:04.173029 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:04 crc kubenswrapper[4953]: I1203 00:08:04.173045 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:04Z","lastTransitionTime":"2025-12-03T00:08:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:04 crc kubenswrapper[4953]: I1203 00:08:04.179602 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f94f8c51-c665-4ca3-bf17-7a04ea5c7ade\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72c5cd7dc1c3a8adfedab6ba43a59d80b3efe65f2bb56be95e76cc86a409dcf3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2810ffc1b141fa7da811406053db9d72c708edf5f2bb1dd75e4d52ab35fd24c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c67b4277b20a9f1287f50ef82d0785c9d02543a8e7ed26e0413eae332b866db7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://258d78e1ad9d3dcda6976909ae3fe71aecead5b995fc66b3fd6e2722d80b5ab8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:06:51Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:08:04Z is after 2025-08-24T17:21:41Z" Dec 03 00:08:04 crc kubenswrapper[4953]: E1203 00:08:04.186593 4953 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:08:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:08:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:08:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:08:04Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:08:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:08:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:08:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:08:04Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"eb4ae5a6-0f6f-49ef-a027-1cdee9d3c536\\\",\\\"systemUUID\\\":\\\"4c1f3d44-3d3d-425d-993f-471223206a89\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:08:04Z is after 2025-08-24T17:21:41Z" Dec 03 00:08:04 crc kubenswrapper[4953]: I1203 00:08:04.191021 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:04 crc kubenswrapper[4953]: I1203 00:08:04.191066 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:04 crc kubenswrapper[4953]: I1203 00:08:04.191080 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:04 crc kubenswrapper[4953]: I1203 00:08:04.191097 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:04 crc kubenswrapper[4953]: I1203 00:08:04.191111 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:04Z","lastTransitionTime":"2025-12-03T00:08:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:04 crc kubenswrapper[4953]: I1203 00:08:04.195496 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5652d0f3-3fd7-4af6-8941-9119e1177641\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2030aef5b7170991268fa5dc7e7ed8ce1fbe3a2765165a10c25251426575330f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ab6052f1abf17225155ce1bd86afa742e6e78f507b7deedd7161cd8fba9120b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://650b6c79cda2e1768ba1d8763db799e210ce2748893512068d25e60f821937ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d1d60162f7f6885d4f51b4ab44163cfae6d2d04355b1dd710e26b022450005df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d1d60162f7f6885d4f51b4ab44163cfae6d2d04355b1dd710e26b022450005df\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:06:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:06:51Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:08:04Z is after 2025-08-24T17:21:41Z" Dec 03 00:08:04 crc kubenswrapper[4953]: E1203 00:08:04.203867 4953 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:08:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:08:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:08:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:08:04Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:08:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:08:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:08:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:08:04Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"eb4ae5a6-0f6f-49ef-a027-1cdee9d3c536\\\",\\\"systemUUID\\\":\\\"4c1f3d44-3d3d-425d-993f-471223206a89\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:08:04Z is after 2025-08-24T17:21:41Z" Dec 03 00:08:04 crc kubenswrapper[4953]: I1203 00:08:04.207514 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:04 crc kubenswrapper[4953]: I1203 00:08:04.207565 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:04 crc kubenswrapper[4953]: I1203 00:08:04.207583 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:04 crc kubenswrapper[4953]: I1203 00:08:04.207609 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:04 crc kubenswrapper[4953]: I1203 00:08:04.207625 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:04Z","lastTransitionTime":"2025-12-03T00:08:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:04 crc kubenswrapper[4953]: E1203 00:08:04.222500 4953 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:08:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:08:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:08:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:08:04Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:08:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:08:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:08:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:08:04Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"eb4ae5a6-0f6f-49ef-a027-1cdee9d3c536\\\",\\\"systemUUID\\\":\\\"4c1f3d44-3d3d-425d-993f-471223206a89\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:08:04Z is after 2025-08-24T17:21:41Z" Dec 03 00:08:04 crc kubenswrapper[4953]: E1203 00:08:04.222664 4953 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 03 00:08:04 crc kubenswrapper[4953]: I1203 00:08:04.224722 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:04 crc kubenswrapper[4953]: I1203 00:08:04.224754 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:04 crc kubenswrapper[4953]: I1203 00:08:04.224764 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:04 crc kubenswrapper[4953]: I1203 00:08:04.224782 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:04 crc kubenswrapper[4953]: I1203 00:08:04.224795 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:04Z","lastTransitionTime":"2025-12-03T00:08:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:04 crc kubenswrapper[4953]: I1203 00:08:04.328025 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:04 crc kubenswrapper[4953]: I1203 00:08:04.328102 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:04 crc kubenswrapper[4953]: I1203 00:08:04.328111 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:04 crc kubenswrapper[4953]: I1203 00:08:04.328129 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:04 crc kubenswrapper[4953]: I1203 00:08:04.328139 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:04Z","lastTransitionTime":"2025-12-03T00:08:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:04 crc kubenswrapper[4953]: I1203 00:08:04.347340 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 00:08:04 crc kubenswrapper[4953]: I1203 00:08:04.347392 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 00:08:04 crc kubenswrapper[4953]: I1203 00:08:04.347344 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 00:08:04 crc kubenswrapper[4953]: E1203 00:08:04.347478 4953 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 00:08:04 crc kubenswrapper[4953]: E1203 00:08:04.347569 4953 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 00:08:04 crc kubenswrapper[4953]: E1203 00:08:04.347649 4953 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 00:08:04 crc kubenswrapper[4953]: I1203 00:08:04.430692 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:04 crc kubenswrapper[4953]: I1203 00:08:04.430755 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:04 crc kubenswrapper[4953]: I1203 00:08:04.430764 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:04 crc kubenswrapper[4953]: I1203 00:08:04.430788 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:04 crc kubenswrapper[4953]: I1203 00:08:04.430799 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:04Z","lastTransitionTime":"2025-12-03T00:08:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:04 crc kubenswrapper[4953]: I1203 00:08:04.533778 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:04 crc kubenswrapper[4953]: I1203 00:08:04.533836 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:04 crc kubenswrapper[4953]: I1203 00:08:04.533882 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:04 crc kubenswrapper[4953]: I1203 00:08:04.533907 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:04 crc kubenswrapper[4953]: I1203 00:08:04.533924 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:04Z","lastTransitionTime":"2025-12-03T00:08:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:04 crc kubenswrapper[4953]: I1203 00:08:04.637059 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:04 crc kubenswrapper[4953]: I1203 00:08:04.637138 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:04 crc kubenswrapper[4953]: I1203 00:08:04.637158 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:04 crc kubenswrapper[4953]: I1203 00:08:04.637190 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:04 crc kubenswrapper[4953]: I1203 00:08:04.637210 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:04Z","lastTransitionTime":"2025-12-03T00:08:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:04 crc kubenswrapper[4953]: I1203 00:08:04.740559 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:04 crc kubenswrapper[4953]: I1203 00:08:04.740613 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:04 crc kubenswrapper[4953]: I1203 00:08:04.740630 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:04 crc kubenswrapper[4953]: I1203 00:08:04.740654 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:04 crc kubenswrapper[4953]: I1203 00:08:04.740673 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:04Z","lastTransitionTime":"2025-12-03T00:08:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:04 crc kubenswrapper[4953]: I1203 00:08:04.843533 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:04 crc kubenswrapper[4953]: I1203 00:08:04.843682 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:04 crc kubenswrapper[4953]: I1203 00:08:04.843701 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:04 crc kubenswrapper[4953]: I1203 00:08:04.843724 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:04 crc kubenswrapper[4953]: I1203 00:08:04.844025 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:04Z","lastTransitionTime":"2025-12-03T00:08:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:04 crc kubenswrapper[4953]: I1203 00:08:04.947764 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:04 crc kubenswrapper[4953]: I1203 00:08:04.947856 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:04 crc kubenswrapper[4953]: I1203 00:08:04.947892 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:04 crc kubenswrapper[4953]: I1203 00:08:04.947925 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:04 crc kubenswrapper[4953]: I1203 00:08:04.947947 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:04Z","lastTransitionTime":"2025-12-03T00:08:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:05 crc kubenswrapper[4953]: I1203 00:08:05.051823 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:05 crc kubenswrapper[4953]: I1203 00:08:05.051946 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:05 crc kubenswrapper[4953]: I1203 00:08:05.052006 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:05 crc kubenswrapper[4953]: I1203 00:08:05.052043 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:05 crc kubenswrapper[4953]: I1203 00:08:05.052074 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:05Z","lastTransitionTime":"2025-12-03T00:08:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:05 crc kubenswrapper[4953]: I1203 00:08:05.155139 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:05 crc kubenswrapper[4953]: I1203 00:08:05.155224 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:05 crc kubenswrapper[4953]: I1203 00:08:05.155247 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:05 crc kubenswrapper[4953]: I1203 00:08:05.155275 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:05 crc kubenswrapper[4953]: I1203 00:08:05.155302 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:05Z","lastTransitionTime":"2025-12-03T00:08:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:05 crc kubenswrapper[4953]: I1203 00:08:05.258455 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:05 crc kubenswrapper[4953]: I1203 00:08:05.258517 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:05 crc kubenswrapper[4953]: I1203 00:08:05.258537 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:05 crc kubenswrapper[4953]: I1203 00:08:05.258566 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:05 crc kubenswrapper[4953]: I1203 00:08:05.258583 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:05Z","lastTransitionTime":"2025-12-03T00:08:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:05 crc kubenswrapper[4953]: I1203 00:08:05.348139 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xshk4" Dec 03 00:08:05 crc kubenswrapper[4953]: E1203 00:08:05.348356 4953 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xshk4" podUID="bd03a758-0ca9-47d8-86f7-18b2d610df43" Dec 03 00:08:05 crc kubenswrapper[4953]: I1203 00:08:05.361561 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:05 crc kubenswrapper[4953]: I1203 00:08:05.361628 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:05 crc kubenswrapper[4953]: I1203 00:08:05.361663 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:05 crc kubenswrapper[4953]: I1203 00:08:05.361699 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:05 crc kubenswrapper[4953]: I1203 00:08:05.361728 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:05Z","lastTransitionTime":"2025-12-03T00:08:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:05 crc kubenswrapper[4953]: I1203 00:08:05.465117 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:05 crc kubenswrapper[4953]: I1203 00:08:05.465182 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:05 crc kubenswrapper[4953]: I1203 00:08:05.465194 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:05 crc kubenswrapper[4953]: I1203 00:08:05.465212 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:05 crc kubenswrapper[4953]: I1203 00:08:05.465225 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:05Z","lastTransitionTime":"2025-12-03T00:08:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:05 crc kubenswrapper[4953]: I1203 00:08:05.569367 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:05 crc kubenswrapper[4953]: I1203 00:08:05.569435 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:05 crc kubenswrapper[4953]: I1203 00:08:05.569453 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:05 crc kubenswrapper[4953]: I1203 00:08:05.569477 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:05 crc kubenswrapper[4953]: I1203 00:08:05.569495 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:05Z","lastTransitionTime":"2025-12-03T00:08:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:05 crc kubenswrapper[4953]: I1203 00:08:05.673237 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:05 crc kubenswrapper[4953]: I1203 00:08:05.673305 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:05 crc kubenswrapper[4953]: I1203 00:08:05.673325 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:05 crc kubenswrapper[4953]: I1203 00:08:05.673350 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:05 crc kubenswrapper[4953]: I1203 00:08:05.673367 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:05Z","lastTransitionTime":"2025-12-03T00:08:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:05 crc kubenswrapper[4953]: I1203 00:08:05.776342 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:05 crc kubenswrapper[4953]: I1203 00:08:05.776390 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:05 crc kubenswrapper[4953]: I1203 00:08:05.776407 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:05 crc kubenswrapper[4953]: I1203 00:08:05.776434 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:05 crc kubenswrapper[4953]: I1203 00:08:05.776460 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:05Z","lastTransitionTime":"2025-12-03T00:08:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:05 crc kubenswrapper[4953]: I1203 00:08:05.879136 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:05 crc kubenswrapper[4953]: I1203 00:08:05.879198 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:05 crc kubenswrapper[4953]: I1203 00:08:05.879216 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:05 crc kubenswrapper[4953]: I1203 00:08:05.879239 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:05 crc kubenswrapper[4953]: I1203 00:08:05.879295 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:05Z","lastTransitionTime":"2025-12-03T00:08:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:05 crc kubenswrapper[4953]: I1203 00:08:05.982516 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:05 crc kubenswrapper[4953]: I1203 00:08:05.982602 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:05 crc kubenswrapper[4953]: I1203 00:08:05.982619 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:05 crc kubenswrapper[4953]: I1203 00:08:05.982643 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:05 crc kubenswrapper[4953]: I1203 00:08:05.982659 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:05Z","lastTransitionTime":"2025-12-03T00:08:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:06 crc kubenswrapper[4953]: I1203 00:08:06.087483 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:06 crc kubenswrapper[4953]: I1203 00:08:06.087556 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:06 crc kubenswrapper[4953]: I1203 00:08:06.087579 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:06 crc kubenswrapper[4953]: I1203 00:08:06.087607 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:06 crc kubenswrapper[4953]: I1203 00:08:06.087630 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:06Z","lastTransitionTime":"2025-12-03T00:08:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:06 crc kubenswrapper[4953]: I1203 00:08:06.190630 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:06 crc kubenswrapper[4953]: I1203 00:08:06.190755 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:06 crc kubenswrapper[4953]: I1203 00:08:06.190782 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:06 crc kubenswrapper[4953]: I1203 00:08:06.190812 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:06 crc kubenswrapper[4953]: I1203 00:08:06.190836 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:06Z","lastTransitionTime":"2025-12-03T00:08:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:06 crc kubenswrapper[4953]: I1203 00:08:06.293217 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:06 crc kubenswrapper[4953]: I1203 00:08:06.293250 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:06 crc kubenswrapper[4953]: I1203 00:08:06.293268 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:06 crc kubenswrapper[4953]: I1203 00:08:06.293285 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:06 crc kubenswrapper[4953]: I1203 00:08:06.293296 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:06Z","lastTransitionTime":"2025-12-03T00:08:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:06 crc kubenswrapper[4953]: I1203 00:08:06.347398 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 00:08:06 crc kubenswrapper[4953]: E1203 00:08:06.347616 4953 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 00:08:06 crc kubenswrapper[4953]: I1203 00:08:06.347935 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 00:08:06 crc kubenswrapper[4953]: I1203 00:08:06.348011 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 00:08:06 crc kubenswrapper[4953]: E1203 00:08:06.348200 4953 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 00:08:06 crc kubenswrapper[4953]: E1203 00:08:06.348303 4953 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 00:08:06 crc kubenswrapper[4953]: I1203 00:08:06.396386 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:06 crc kubenswrapper[4953]: I1203 00:08:06.396466 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:06 crc kubenswrapper[4953]: I1203 00:08:06.396488 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:06 crc kubenswrapper[4953]: I1203 00:08:06.396518 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:06 crc kubenswrapper[4953]: I1203 00:08:06.396539 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:06Z","lastTransitionTime":"2025-12-03T00:08:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:06 crc kubenswrapper[4953]: I1203 00:08:06.501438 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:06 crc kubenswrapper[4953]: I1203 00:08:06.501505 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:06 crc kubenswrapper[4953]: I1203 00:08:06.501526 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:06 crc kubenswrapper[4953]: I1203 00:08:06.501554 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:06 crc kubenswrapper[4953]: I1203 00:08:06.501575 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:06Z","lastTransitionTime":"2025-12-03T00:08:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:06 crc kubenswrapper[4953]: I1203 00:08:06.604611 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:06 crc kubenswrapper[4953]: I1203 00:08:06.604675 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:06 crc kubenswrapper[4953]: I1203 00:08:06.604694 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:06 crc kubenswrapper[4953]: I1203 00:08:06.604723 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:06 crc kubenswrapper[4953]: I1203 00:08:06.604741 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:06Z","lastTransitionTime":"2025-12-03T00:08:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:06 crc kubenswrapper[4953]: I1203 00:08:06.707801 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:06 crc kubenswrapper[4953]: I1203 00:08:06.707890 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:06 crc kubenswrapper[4953]: I1203 00:08:06.707910 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:06 crc kubenswrapper[4953]: I1203 00:08:06.707935 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:06 crc kubenswrapper[4953]: I1203 00:08:06.707954 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:06Z","lastTransitionTime":"2025-12-03T00:08:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:06 crc kubenswrapper[4953]: I1203 00:08:06.811076 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:06 crc kubenswrapper[4953]: I1203 00:08:06.811120 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:06 crc kubenswrapper[4953]: I1203 00:08:06.811131 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:06 crc kubenswrapper[4953]: I1203 00:08:06.811146 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:06 crc kubenswrapper[4953]: I1203 00:08:06.811158 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:06Z","lastTransitionTime":"2025-12-03T00:08:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:06 crc kubenswrapper[4953]: I1203 00:08:06.914294 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:06 crc kubenswrapper[4953]: I1203 00:08:06.914371 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:06 crc kubenswrapper[4953]: I1203 00:08:06.914397 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:06 crc kubenswrapper[4953]: I1203 00:08:06.914423 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:06 crc kubenswrapper[4953]: I1203 00:08:06.914441 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:06Z","lastTransitionTime":"2025-12-03T00:08:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:07 crc kubenswrapper[4953]: I1203 00:08:07.016681 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:07 crc kubenswrapper[4953]: I1203 00:08:07.016745 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:07 crc kubenswrapper[4953]: I1203 00:08:07.016764 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:07 crc kubenswrapper[4953]: I1203 00:08:07.016788 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:07 crc kubenswrapper[4953]: I1203 00:08:07.016806 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:07Z","lastTransitionTime":"2025-12-03T00:08:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:07 crc kubenswrapper[4953]: I1203 00:08:07.120227 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:07 crc kubenswrapper[4953]: I1203 00:08:07.120312 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:07 crc kubenswrapper[4953]: I1203 00:08:07.120339 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:07 crc kubenswrapper[4953]: I1203 00:08:07.120373 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:07 crc kubenswrapper[4953]: I1203 00:08:07.120397 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:07Z","lastTransitionTime":"2025-12-03T00:08:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:07 crc kubenswrapper[4953]: I1203 00:08:07.223681 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:07 crc kubenswrapper[4953]: I1203 00:08:07.223754 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:07 crc kubenswrapper[4953]: I1203 00:08:07.223766 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:07 crc kubenswrapper[4953]: I1203 00:08:07.223784 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:07 crc kubenswrapper[4953]: I1203 00:08:07.223795 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:07Z","lastTransitionTime":"2025-12-03T00:08:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:07 crc kubenswrapper[4953]: I1203 00:08:07.326337 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:07 crc kubenswrapper[4953]: I1203 00:08:07.326405 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:07 crc kubenswrapper[4953]: I1203 00:08:07.326427 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:07 crc kubenswrapper[4953]: I1203 00:08:07.326459 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:07 crc kubenswrapper[4953]: I1203 00:08:07.326481 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:07Z","lastTransitionTime":"2025-12-03T00:08:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:07 crc kubenswrapper[4953]: I1203 00:08:07.347322 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xshk4" Dec 03 00:08:07 crc kubenswrapper[4953]: E1203 00:08:07.347547 4953 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xshk4" podUID="bd03a758-0ca9-47d8-86f7-18b2d610df43" Dec 03 00:08:07 crc kubenswrapper[4953]: I1203 00:08:07.430688 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:07 crc kubenswrapper[4953]: I1203 00:08:07.430775 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:07 crc kubenswrapper[4953]: I1203 00:08:07.430794 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:07 crc kubenswrapper[4953]: I1203 00:08:07.430827 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:07 crc kubenswrapper[4953]: I1203 00:08:07.430851 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:07Z","lastTransitionTime":"2025-12-03T00:08:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:07 crc kubenswrapper[4953]: I1203 00:08:07.534546 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:07 crc kubenswrapper[4953]: I1203 00:08:07.534632 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:07 crc kubenswrapper[4953]: I1203 00:08:07.534664 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:07 crc kubenswrapper[4953]: I1203 00:08:07.534701 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:07 crc kubenswrapper[4953]: I1203 00:08:07.534722 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:07Z","lastTransitionTime":"2025-12-03T00:08:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:07 crc kubenswrapper[4953]: I1203 00:08:07.637709 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:07 crc kubenswrapper[4953]: I1203 00:08:07.637764 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:07 crc kubenswrapper[4953]: I1203 00:08:07.637783 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:07 crc kubenswrapper[4953]: I1203 00:08:07.637804 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:07 crc kubenswrapper[4953]: I1203 00:08:07.637821 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:07Z","lastTransitionTime":"2025-12-03T00:08:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:07 crc kubenswrapper[4953]: I1203 00:08:07.740105 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:07 crc kubenswrapper[4953]: I1203 00:08:07.740169 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:07 crc kubenswrapper[4953]: I1203 00:08:07.740190 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:07 crc kubenswrapper[4953]: I1203 00:08:07.740214 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:07 crc kubenswrapper[4953]: I1203 00:08:07.740238 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:07Z","lastTransitionTime":"2025-12-03T00:08:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:07 crc kubenswrapper[4953]: I1203 00:08:07.842146 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:07 crc kubenswrapper[4953]: I1203 00:08:07.842257 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:07 crc kubenswrapper[4953]: I1203 00:08:07.842277 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:07 crc kubenswrapper[4953]: I1203 00:08:07.842299 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:07 crc kubenswrapper[4953]: I1203 00:08:07.842316 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:07Z","lastTransitionTime":"2025-12-03T00:08:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:07 crc kubenswrapper[4953]: I1203 00:08:07.945608 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:07 crc kubenswrapper[4953]: I1203 00:08:07.945674 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:07 crc kubenswrapper[4953]: I1203 00:08:07.945696 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:07 crc kubenswrapper[4953]: I1203 00:08:07.945723 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:07 crc kubenswrapper[4953]: I1203 00:08:07.945744 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:07Z","lastTransitionTime":"2025-12-03T00:08:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:08 crc kubenswrapper[4953]: I1203 00:08:08.049186 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:08 crc kubenswrapper[4953]: I1203 00:08:08.049316 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:08 crc kubenswrapper[4953]: I1203 00:08:08.049351 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:08 crc kubenswrapper[4953]: I1203 00:08:08.049377 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:08 crc kubenswrapper[4953]: I1203 00:08:08.049397 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:08Z","lastTransitionTime":"2025-12-03T00:08:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:08 crc kubenswrapper[4953]: I1203 00:08:08.152401 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:08 crc kubenswrapper[4953]: I1203 00:08:08.152476 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:08 crc kubenswrapper[4953]: I1203 00:08:08.152489 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:08 crc kubenswrapper[4953]: I1203 00:08:08.152512 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:08 crc kubenswrapper[4953]: I1203 00:08:08.152529 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:08Z","lastTransitionTime":"2025-12-03T00:08:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:08 crc kubenswrapper[4953]: I1203 00:08:08.256575 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:08 crc kubenswrapper[4953]: I1203 00:08:08.256631 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:08 crc kubenswrapper[4953]: I1203 00:08:08.256641 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:08 crc kubenswrapper[4953]: I1203 00:08:08.256668 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:08 crc kubenswrapper[4953]: I1203 00:08:08.256683 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:08Z","lastTransitionTime":"2025-12-03T00:08:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:08 crc kubenswrapper[4953]: I1203 00:08:08.347305 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 00:08:08 crc kubenswrapper[4953]: E1203 00:08:08.347512 4953 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 00:08:08 crc kubenswrapper[4953]: I1203 00:08:08.347824 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 00:08:08 crc kubenswrapper[4953]: E1203 00:08:08.347917 4953 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 00:08:08 crc kubenswrapper[4953]: I1203 00:08:08.348124 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 00:08:08 crc kubenswrapper[4953]: E1203 00:08:08.348249 4953 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 00:08:08 crc kubenswrapper[4953]: I1203 00:08:08.359688 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:08 crc kubenswrapper[4953]: I1203 00:08:08.359748 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:08 crc kubenswrapper[4953]: I1203 00:08:08.359761 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:08 crc kubenswrapper[4953]: I1203 00:08:08.359784 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:08 crc kubenswrapper[4953]: I1203 00:08:08.359797 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:08Z","lastTransitionTime":"2025-12-03T00:08:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:08 crc kubenswrapper[4953]: I1203 00:08:08.463645 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:08 crc kubenswrapper[4953]: I1203 00:08:08.463766 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:08 crc kubenswrapper[4953]: I1203 00:08:08.463780 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:08 crc kubenswrapper[4953]: I1203 00:08:08.463806 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:08 crc kubenswrapper[4953]: I1203 00:08:08.463821 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:08Z","lastTransitionTime":"2025-12-03T00:08:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:08 crc kubenswrapper[4953]: I1203 00:08:08.570644 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:08 crc kubenswrapper[4953]: I1203 00:08:08.570686 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:08 crc kubenswrapper[4953]: I1203 00:08:08.570695 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:08 crc kubenswrapper[4953]: I1203 00:08:08.570710 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:08 crc kubenswrapper[4953]: I1203 00:08:08.570721 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:08Z","lastTransitionTime":"2025-12-03T00:08:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:08 crc kubenswrapper[4953]: I1203 00:08:08.673603 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:08 crc kubenswrapper[4953]: I1203 00:08:08.673639 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:08 crc kubenswrapper[4953]: I1203 00:08:08.673647 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:08 crc kubenswrapper[4953]: I1203 00:08:08.673659 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:08 crc kubenswrapper[4953]: I1203 00:08:08.673668 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:08Z","lastTransitionTime":"2025-12-03T00:08:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:08 crc kubenswrapper[4953]: I1203 00:08:08.775679 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:08 crc kubenswrapper[4953]: I1203 00:08:08.775750 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:08 crc kubenswrapper[4953]: I1203 00:08:08.775774 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:08 crc kubenswrapper[4953]: I1203 00:08:08.775807 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:08 crc kubenswrapper[4953]: I1203 00:08:08.775872 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:08Z","lastTransitionTime":"2025-12-03T00:08:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:08 crc kubenswrapper[4953]: I1203 00:08:08.878539 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:08 crc kubenswrapper[4953]: I1203 00:08:08.878648 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:08 crc kubenswrapper[4953]: I1203 00:08:08.878670 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:08 crc kubenswrapper[4953]: I1203 00:08:08.878696 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:08 crc kubenswrapper[4953]: I1203 00:08:08.878713 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:08Z","lastTransitionTime":"2025-12-03T00:08:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:08 crc kubenswrapper[4953]: I1203 00:08:08.981823 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:08 crc kubenswrapper[4953]: I1203 00:08:08.982244 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:08 crc kubenswrapper[4953]: I1203 00:08:08.982412 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:08 crc kubenswrapper[4953]: I1203 00:08:08.982556 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:08 crc kubenswrapper[4953]: I1203 00:08:08.982680 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:08Z","lastTransitionTime":"2025-12-03T00:08:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:09 crc kubenswrapper[4953]: I1203 00:08:09.085443 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:09 crc kubenswrapper[4953]: I1203 00:08:09.085490 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:09 crc kubenswrapper[4953]: I1203 00:08:09.085508 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:09 crc kubenswrapper[4953]: I1203 00:08:09.085531 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:09 crc kubenswrapper[4953]: I1203 00:08:09.085548 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:09Z","lastTransitionTime":"2025-12-03T00:08:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:09 crc kubenswrapper[4953]: I1203 00:08:09.188193 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:09 crc kubenswrapper[4953]: I1203 00:08:09.188698 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:09 crc kubenswrapper[4953]: I1203 00:08:09.188790 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:09 crc kubenswrapper[4953]: I1203 00:08:09.188868 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:09 crc kubenswrapper[4953]: I1203 00:08:09.188925 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:09Z","lastTransitionTime":"2025-12-03T00:08:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:09 crc kubenswrapper[4953]: I1203 00:08:09.291610 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:09 crc kubenswrapper[4953]: I1203 00:08:09.291952 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:09 crc kubenswrapper[4953]: I1203 00:08:09.292214 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:09 crc kubenswrapper[4953]: I1203 00:08:09.292365 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:09 crc kubenswrapper[4953]: I1203 00:08:09.292521 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:09Z","lastTransitionTime":"2025-12-03T00:08:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:09 crc kubenswrapper[4953]: I1203 00:08:09.348166 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xshk4" Dec 03 00:08:09 crc kubenswrapper[4953]: E1203 00:08:09.348668 4953 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xshk4" podUID="bd03a758-0ca9-47d8-86f7-18b2d610df43" Dec 03 00:08:09 crc kubenswrapper[4953]: I1203 00:08:09.396045 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:09 crc kubenswrapper[4953]: I1203 00:08:09.396098 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:09 crc kubenswrapper[4953]: I1203 00:08:09.396154 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:09 crc kubenswrapper[4953]: I1203 00:08:09.396177 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:09 crc kubenswrapper[4953]: I1203 00:08:09.396197 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:09Z","lastTransitionTime":"2025-12-03T00:08:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:09 crc kubenswrapper[4953]: I1203 00:08:09.499880 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:09 crc kubenswrapper[4953]: I1203 00:08:09.499939 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:09 crc kubenswrapper[4953]: I1203 00:08:09.500017 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:09 crc kubenswrapper[4953]: I1203 00:08:09.500058 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:09 crc kubenswrapper[4953]: I1203 00:08:09.500076 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:09Z","lastTransitionTime":"2025-12-03T00:08:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:09 crc kubenswrapper[4953]: I1203 00:08:09.606494 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:09 crc kubenswrapper[4953]: I1203 00:08:09.606568 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:09 crc kubenswrapper[4953]: I1203 00:08:09.606595 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:09 crc kubenswrapper[4953]: I1203 00:08:09.606626 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:09 crc kubenswrapper[4953]: I1203 00:08:09.606648 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:09Z","lastTransitionTime":"2025-12-03T00:08:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:09 crc kubenswrapper[4953]: I1203 00:08:09.709612 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:09 crc kubenswrapper[4953]: I1203 00:08:09.709722 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:09 crc kubenswrapper[4953]: I1203 00:08:09.709741 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:09 crc kubenswrapper[4953]: I1203 00:08:09.709769 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:09 crc kubenswrapper[4953]: I1203 00:08:09.709792 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:09Z","lastTransitionTime":"2025-12-03T00:08:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:09 crc kubenswrapper[4953]: I1203 00:08:09.812199 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:09 crc kubenswrapper[4953]: I1203 00:08:09.812288 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:09 crc kubenswrapper[4953]: I1203 00:08:09.812370 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:09 crc kubenswrapper[4953]: I1203 00:08:09.812411 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:09 crc kubenswrapper[4953]: I1203 00:08:09.812438 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:09Z","lastTransitionTime":"2025-12-03T00:08:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:09 crc kubenswrapper[4953]: I1203 00:08:09.915940 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:09 crc kubenswrapper[4953]: I1203 00:08:09.916049 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:09 crc kubenswrapper[4953]: I1203 00:08:09.916075 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:09 crc kubenswrapper[4953]: I1203 00:08:09.916105 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:09 crc kubenswrapper[4953]: I1203 00:08:09.916128 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:09Z","lastTransitionTime":"2025-12-03T00:08:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:10 crc kubenswrapper[4953]: I1203 00:08:10.020041 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:10 crc kubenswrapper[4953]: I1203 00:08:10.020109 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:10 crc kubenswrapper[4953]: I1203 00:08:10.020130 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:10 crc kubenswrapper[4953]: I1203 00:08:10.020169 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:10 crc kubenswrapper[4953]: I1203 00:08:10.020194 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:10Z","lastTransitionTime":"2025-12-03T00:08:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:10 crc kubenswrapper[4953]: I1203 00:08:10.123551 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:10 crc kubenswrapper[4953]: I1203 00:08:10.123629 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:10 crc kubenswrapper[4953]: I1203 00:08:10.123654 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:10 crc kubenswrapper[4953]: I1203 00:08:10.123686 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:10 crc kubenswrapper[4953]: I1203 00:08:10.123708 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:10Z","lastTransitionTime":"2025-12-03T00:08:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:10 crc kubenswrapper[4953]: I1203 00:08:10.227074 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:10 crc kubenswrapper[4953]: I1203 00:08:10.227137 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:10 crc kubenswrapper[4953]: I1203 00:08:10.227153 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:10 crc kubenswrapper[4953]: I1203 00:08:10.227178 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:10 crc kubenswrapper[4953]: I1203 00:08:10.227196 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:10Z","lastTransitionTime":"2025-12-03T00:08:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:10 crc kubenswrapper[4953]: I1203 00:08:10.330136 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:10 crc kubenswrapper[4953]: I1203 00:08:10.330190 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:10 crc kubenswrapper[4953]: I1203 00:08:10.330231 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:10 crc kubenswrapper[4953]: I1203 00:08:10.330262 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:10 crc kubenswrapper[4953]: I1203 00:08:10.330283 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:10Z","lastTransitionTime":"2025-12-03T00:08:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:10 crc kubenswrapper[4953]: I1203 00:08:10.347636 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 00:08:10 crc kubenswrapper[4953]: E1203 00:08:10.347831 4953 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 00:08:10 crc kubenswrapper[4953]: I1203 00:08:10.348199 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 00:08:10 crc kubenswrapper[4953]: E1203 00:08:10.348343 4953 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 00:08:10 crc kubenswrapper[4953]: I1203 00:08:10.348388 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 00:08:10 crc kubenswrapper[4953]: E1203 00:08:10.348502 4953 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 00:08:10 crc kubenswrapper[4953]: I1203 00:08:10.432772 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:10 crc kubenswrapper[4953]: I1203 00:08:10.432826 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:10 crc kubenswrapper[4953]: I1203 00:08:10.432837 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:10 crc kubenswrapper[4953]: I1203 00:08:10.432856 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:10 crc kubenswrapper[4953]: I1203 00:08:10.432871 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:10Z","lastTransitionTime":"2025-12-03T00:08:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:10 crc kubenswrapper[4953]: I1203 00:08:10.536122 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:10 crc kubenswrapper[4953]: I1203 00:08:10.536172 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:10 crc kubenswrapper[4953]: I1203 00:08:10.536187 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:10 crc kubenswrapper[4953]: I1203 00:08:10.536207 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:10 crc kubenswrapper[4953]: I1203 00:08:10.536220 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:10Z","lastTransitionTime":"2025-12-03T00:08:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:10 crc kubenswrapper[4953]: I1203 00:08:10.639791 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:10 crc kubenswrapper[4953]: I1203 00:08:10.639948 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:10 crc kubenswrapper[4953]: I1203 00:08:10.640009 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:10 crc kubenswrapper[4953]: I1203 00:08:10.640035 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:10 crc kubenswrapper[4953]: I1203 00:08:10.640052 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:10Z","lastTransitionTime":"2025-12-03T00:08:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:10 crc kubenswrapper[4953]: I1203 00:08:10.743166 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:10 crc kubenswrapper[4953]: I1203 00:08:10.743254 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:10 crc kubenswrapper[4953]: I1203 00:08:10.743286 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:10 crc kubenswrapper[4953]: I1203 00:08:10.743325 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:10 crc kubenswrapper[4953]: I1203 00:08:10.743361 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:10Z","lastTransitionTime":"2025-12-03T00:08:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:10 crc kubenswrapper[4953]: I1203 00:08:10.846074 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:10 crc kubenswrapper[4953]: I1203 00:08:10.846150 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:10 crc kubenswrapper[4953]: I1203 00:08:10.846173 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:10 crc kubenswrapper[4953]: I1203 00:08:10.846198 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:10 crc kubenswrapper[4953]: I1203 00:08:10.846216 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:10Z","lastTransitionTime":"2025-12-03T00:08:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:10 crc kubenswrapper[4953]: I1203 00:08:10.949464 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:10 crc kubenswrapper[4953]: I1203 00:08:10.949522 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:10 crc kubenswrapper[4953]: I1203 00:08:10.949538 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:10 crc kubenswrapper[4953]: I1203 00:08:10.949564 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:10 crc kubenswrapper[4953]: I1203 00:08:10.949580 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:10Z","lastTransitionTime":"2025-12-03T00:08:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:11 crc kubenswrapper[4953]: I1203 00:08:11.059449 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:11 crc kubenswrapper[4953]: I1203 00:08:11.059530 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:11 crc kubenswrapper[4953]: I1203 00:08:11.059554 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:11 crc kubenswrapper[4953]: I1203 00:08:11.059585 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:11 crc kubenswrapper[4953]: I1203 00:08:11.059611 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:11Z","lastTransitionTime":"2025-12-03T00:08:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:11 crc kubenswrapper[4953]: I1203 00:08:11.162568 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:11 crc kubenswrapper[4953]: I1203 00:08:11.162630 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:11 crc kubenswrapper[4953]: I1203 00:08:11.162647 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:11 crc kubenswrapper[4953]: I1203 00:08:11.162672 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:11 crc kubenswrapper[4953]: I1203 00:08:11.162697 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:11Z","lastTransitionTime":"2025-12-03T00:08:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:11 crc kubenswrapper[4953]: I1203 00:08:11.265432 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:11 crc kubenswrapper[4953]: I1203 00:08:11.265838 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:11 crc kubenswrapper[4953]: I1203 00:08:11.265856 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:11 crc kubenswrapper[4953]: I1203 00:08:11.265878 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:11 crc kubenswrapper[4953]: I1203 00:08:11.265899 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:11Z","lastTransitionTime":"2025-12-03T00:08:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:11 crc kubenswrapper[4953]: I1203 00:08:11.347788 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xshk4" Dec 03 00:08:11 crc kubenswrapper[4953]: E1203 00:08:11.348294 4953 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xshk4" podUID="bd03a758-0ca9-47d8-86f7-18b2d610df43" Dec 03 00:08:11 crc kubenswrapper[4953]: I1203 00:08:11.366661 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7cf23d2d-21d4-493d-bdc7-d088aa876304\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4c95eb1f9b0c797376a1b557415b3f956890c971dc77c5257f238b4f32ff9b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ad2a6f5a3296c6a4f53f9da696e3e3f4d6c919f44b380c9b4114021b8ad74b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ad2a6f5a3296c6a4f53f9da696e3e3f4d6c919f44b380c9b4114021b8ad74b6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:06:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:06:51Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:08:11Z is after 2025-08-24T17:21:41Z" Dec 03 00:08:11 crc kubenswrapper[4953]: I1203 00:08:11.368270 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:11 crc kubenswrapper[4953]: I1203 00:08:11.368335 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:11 crc kubenswrapper[4953]: I1203 00:08:11.368359 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:11 crc kubenswrapper[4953]: I1203 00:08:11.368388 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:11 crc kubenswrapper[4953]: I1203 00:08:11.368410 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:11Z","lastTransitionTime":"2025-12-03T00:08:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:11 crc kubenswrapper[4953]: I1203 00:08:11.388037 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:08:11Z is after 2025-08-24T17:21:41Z" Dec 03 00:08:11 crc kubenswrapper[4953]: I1203 00:08:11.409936 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2ed9c7fa96c8951fbc1728fe7a91fefdd999077285621df9c1abedf9ae0c449\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:08:11Z is after 2025-08-24T17:21:41Z" Dec 03 00:08:11 crc kubenswrapper[4953]: I1203 00:08:11.431897 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:10Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:08:11Z is after 2025-08-24T17:21:41Z" Dec 03 00:08:11 crc kubenswrapper[4953]: I1203 00:08:11.457797 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-675qk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"43a4bad8-71f4-441a-8302-aeb526a986af\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:08:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:08:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e701c2ad05014683415ae2fb03b96f3acb38062ba7c5bfc4d0a9e34431695e36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5b5794fef6477aab5deed8f642e8cc50b37eb1fd0cc6f19c003f34ad31c1e3c0\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T00:08:02Z\\\",\\\"message\\\":\\\"2025-12-03T00:07:17+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_fa87bd16-684d-458f-b6b8-b6ada86b0d46\\\\n2025-12-03T00:07:17+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_fa87bd16-684d-458f-b6b8-b6ada86b0d46 to /host/opt/cni/bin/\\\\n2025-12-03T00:07:17Z [verbose] multus-daemon started\\\\n2025-12-03T00:07:17Z [verbose] Readiness Indicator file check\\\\n2025-12-03T00:08:02Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T00:07:16Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:08:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gdrw4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:07:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-675qk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:08:11Z is after 2025-08-24T17:21:41Z" Dec 03 00:08:11 crc kubenswrapper[4953]: I1203 00:08:11.471043 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:11 crc kubenswrapper[4953]: I1203 00:08:11.471100 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:11 crc kubenswrapper[4953]: I1203 00:08:11.471117 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:11 crc kubenswrapper[4953]: I1203 00:08:11.471139 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:11 crc kubenswrapper[4953]: I1203 00:08:11.471159 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:11Z","lastTransitionTime":"2025-12-03T00:08:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:11 crc kubenswrapper[4953]: I1203 00:08:11.477365 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f94f8c51-c665-4ca3-bf17-7a04ea5c7ade\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72c5cd7dc1c3a8adfedab6ba43a59d80b3efe65f2bb56be95e76cc86a409dcf3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2810ffc1b141fa7da811406053db9d72c708edf5f2bb1dd75e4d52ab35fd24c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c67b4277b20a9f1287f50ef82d0785c9d02543a8e7ed26e0413eae332b866db7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://258d78e1ad9d3dcda6976909ae3fe71aecead5b995fc66b3fd6e2722d80b5ab8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:06:51Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:08:11Z is after 2025-08-24T17:21:41Z" Dec 03 00:08:11 crc kubenswrapper[4953]: I1203 00:08:11.499753 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5652d0f3-3fd7-4af6-8941-9119e1177641\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2030aef5b7170991268fa5dc7e7ed8ce1fbe3a2765165a10c25251426575330f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ab6052f1abf17225155ce1bd86afa742e6e78f507b7deedd7161cd8fba9120b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://650b6c79cda2e1768ba1d8763db799e210ce2748893512068d25e60f821937ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d1d60162f7f6885d4f51b4ab44163cfae6d2d04355b1dd710e26b022450005df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d1d60162f7f6885d4f51b4ab44163cfae6d2d04355b1dd710e26b022450005df\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:06:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:06:51Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:08:11Z is after 2025-08-24T17:21:41Z" Dec 03 00:08:11 crc kubenswrapper[4953]: I1203 00:08:11.519917 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09657a9f448a4c92a759c73a0a892076d7eb46393e6fd565e5f73f34c71aac23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5af5eac4a58870f63497f99b503cfa5fb922278095e0332ec50376925ecf2109\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:08:11Z is after 2025-08-24T17:21:41Z" Dec 03 00:08:11 crc kubenswrapper[4953]: I1203 00:08:11.542613 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-wj6rz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"96aa67a0-15dd-420d-a1d5-ed239d7df1fb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01815cfd3f58c3f3f5f2a894c1790e0b69c5060c48dc0b1deff050b6ad7af916\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kf62m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b12bfb5bdf9fd4f2c50a335fc82df6ee0440eaf8741897da03aa2dae368cc6b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b12bfb5bdf9fd4f2c50a335fc82df6ee0440eaf8741897da03aa2dae368cc6b4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:07:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kf62m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83aaa5fa28110bc0ef05aff75cd1442bc9bbdd33b822ae2ef22d4e4ab83238bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://83aaa5fa28110bc0ef05aff75cd1442bc9bbdd33b822ae2ef22d4e4ab83238bc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:07:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:07:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kf62m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ab6213a33a19fedebce6d74a0ebbb3b3a2de5687504e21cc24093cc42ade635\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ab6213a33a19fedebce6d74a0ebbb3b3a2de5687504e21cc24093cc42ade635\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:07:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:07:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kf62m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://de480488843afa6b8dede1e1b42b5181dca1993e2f39dc1b461a18a57e3668e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://de480488843afa6b8dede1e1b42b5181dca1993e2f39dc1b461a18a57e3668e1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:07:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kf62m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://97be4ba692bc61078d4b0ba39fb0c12772548c6c9251ff3356ea702f151223cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://97be4ba692bc61078d4b0ba39fb0c12772548c6c9251ff3356ea702f151223cd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:07:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kf62m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1a1f009fcbd8af9217db1dc97f662719f8c1db232df0e2804e44c881abb1df2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b1a1f009fcbd8af9217db1dc97f662719f8c1db232df0e2804e44c881abb1df2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:07:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:07:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kf62m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:07:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-wj6rz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:08:11Z is after 2025-08-24T17:21:41Z" Dec 03 00:08:11 crc kubenswrapper[4953]: I1203 00:08:11.562422 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-kl5s9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d85e5c97-16ad-4bd7-9e73-b7f12678be97\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c21df91a9e4e3165686882b2eb559498224063dc633c3a63f958598108eca29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j24lb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:07:16Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-kl5s9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:08:11Z is after 2025-08-24T17:21:41Z" Dec 03 00:08:11 crc kubenswrapper[4953]: I1203 00:08:11.574362 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:11 crc kubenswrapper[4953]: I1203 00:08:11.574403 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:11 crc kubenswrapper[4953]: I1203 00:08:11.574413 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:11 crc kubenswrapper[4953]: I1203 00:08:11.574428 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:11 crc kubenswrapper[4953]: I1203 00:08:11.574437 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:11Z","lastTransitionTime":"2025-12-03T00:08:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:11 crc kubenswrapper[4953]: I1203 00:08:11.583126 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cf821840-5353-4620-8f73-99b56b371ab0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cae5e2994fcc24dc9b2ea73e79572615b3311edc42642b7f278fb84123989b23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://29e4b9d1b70a93fb133e0759ad148ae133766a66efeae3ff06a21dfae0e805ff\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dec43536772aa7c3f5a662fa7f3bd73b1cae2791ad454f85792287828d0ffeed\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://034e0bac6892d361700b8f33d6427b8a55e2bd3f89b14ea77a374c6be94d25e2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://761973393a93c59c23e69f41b8ac0d16abc35abd61ecbc6d51a42d94cd664233\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T00:07:09Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 00:07:03.713374 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 00:07:03.715052 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1528563048/tls.crt::/tmp/serving-cert-1528563048/tls.key\\\\\\\"\\\\nI1203 00:07:09.300052 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 00:07:09.304192 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 00:07:09.304255 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 00:07:09.304309 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 00:07:09.304323 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 00:07:09.313563 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 00:07:09.313597 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 00:07:09.313603 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 00:07:09.313609 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 00:07:09.313613 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 00:07:09.313618 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 00:07:09.313622 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 00:07:09.313709 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 00:07:09.327347 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T00:06:53Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c143d4232fb8bf7ef5999a664e431a00ab0db030f4d31a28d0bb21e00e38581\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:53Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72af6c27296eae2e275d455795c68f0d272869d63642b8353b17d466d0c901cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://72af6c27296eae2e275d455795c68f0d272869d63642b8353b17d466d0c901cf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:06:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:06:51Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:08:11Z is after 2025-08-24T17:21:41Z" Dec 03 00:08:11 crc kubenswrapper[4953]: I1203 00:08:11.602001 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:10Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:08:11Z is after 2025-08-24T17:21:41Z" Dec 03 00:08:11 crc kubenswrapper[4953]: I1203 00:08:11.615586 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce8b414e1fc22eb30decb6a6411b672e7f443055cb89ebb029fbf6ffc1d5dca7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:08:11Z is after 2025-08-24T17:21:41Z" Dec 03 00:08:11 crc kubenswrapper[4953]: I1203 00:08:11.629138 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-6dfbn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53ffb1e9-9a2e-4869-80d7-776f4aa7ef8b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ab4205c0d38ad6a0c32077eadb1fd5fd3a279d9c8b8a2ffbe693ec2c4018370c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ct26c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7df84405eda520ad5a22d35750edf07ef3c3977a756ad5739c54f91cba9da7b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ct26c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:07:16Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-6dfbn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:08:11Z is after 2025-08-24T17:21:41Z" Dec 03 00:08:11 crc kubenswrapper[4953]: I1203 00:08:11.653940 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-txffx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8816ebee-c735-4b8d-b1ec-65349707b347\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:16Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:16Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4b85224419a4075acb340e4b11095f2adb5c316e285afa4e19295c2181c0d24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0f6f645a58993f052a6f365bb808f62d99f7fea74886c5cc022b2d86112321a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a61b951a9cbab2c48c0f2df8351f97d7ff71bf3674f641fe5eac30155232b224\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd5f428de229749485267e81d34e31b708728b52474032e88460e54747f21e11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://00a2ea4a9bb870fce30e7fa845f787f184e7a19e42c95511655ea8c5f84a26d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc0a5d4cc9d614d8492626a1b114bdc52f510eb5965d842c02c083f08d4849a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dcb805f9ce285d3841e93d588fe7f3d42232382366d42b75bcfb129aefa91c1f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dcb805f9ce285d3841e93d588fe7f3d42232382366d42b75bcfb129aefa91c1f\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T00:07:44Z\\\",\\\"message\\\":\\\"mers/externalversions/factory.go:140\\\\nI1203 00:07:44.043157 6594 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1203 00:07:44.043232 6594 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 00:07:44.043450 6594 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 00:07:44.043811 6594 reflector.go:311] Stopping reflector *v1.NetworkAttachmentDefinition (0s) from github.com/k8snetworkplumbingwg/network-attachment-definition-client/pkg/client/informers/externalversions/factory.go:117\\\\nI1203 00:07:44.044028 6594 reflector.go:311] Stopping reflector *v1.AdminPolicyBasedExternalRoute (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140\\\\nI1203 00:07:44.044197 6594 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI1203 00:07:44.044257 6594 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T00:07:43Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-txffx_openshift-ovn-kubernetes(8816ebee-c735-4b8d-b1ec-65349707b347)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7deadc8e135ff1bc7c5a2f093e489bb5b4b6f166bc874c6acc2a841c5397f11e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af4c7764fca87144d01f3738ea010308744ccf50c54b7c471ccc213112de00d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af4c7764fca87144d01f3738ea010308744ccf50c54b7c471ccc213112de00d5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:07:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:07:16Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-txffx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:08:11Z is after 2025-08-24T17:21:41Z" Dec 03 00:08:11 crc kubenswrapper[4953]: I1203 00:08:11.671566 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5cmlc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7e11a1d3-ecee-4bf0-87bf-085fab97b73d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d42390b9c53efc63add8ddcebf03c976f5e5448eaa2c34e076b2dc90dbac60fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6x2sc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:07:18Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5cmlc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:08:11Z is after 2025-08-24T17:21:41Z" Dec 03 00:08:11 crc kubenswrapper[4953]: I1203 00:08:11.677368 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:11 crc kubenswrapper[4953]: I1203 00:08:11.677446 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:11 crc kubenswrapper[4953]: I1203 00:08:11.677473 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:11 crc kubenswrapper[4953]: I1203 00:08:11.677506 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:11 crc kubenswrapper[4953]: I1203 00:08:11.677528 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:11Z","lastTransitionTime":"2025-12-03T00:08:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:11 crc kubenswrapper[4953]: I1203 00:08:11.690250 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-chmnz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a12fe6d-d25a-4b08-90da-cc9e510980a7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://711ab839cb9c0a90554dadbd45d810fc78aac09739340482fef8de8743b77283\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c982g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f802d5628a8cc32618044b9383211d5a492c56060d427f953b7793a1064029e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c982g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:07:29Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-chmnz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:08:11Z is after 2025-08-24T17:21:41Z" Dec 03 00:08:11 crc kubenswrapper[4953]: I1203 00:08:11.706602 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-xshk4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bd03a758-0ca9-47d8-86f7-18b2d610df43\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47gfh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47gfh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:07:31Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-xshk4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:08:11Z is after 2025-08-24T17:21:41Z" Dec 03 00:08:11 crc kubenswrapper[4953]: I1203 00:08:11.781120 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:11 crc kubenswrapper[4953]: I1203 00:08:11.781233 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:11 crc kubenswrapper[4953]: I1203 00:08:11.781258 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:11 crc kubenswrapper[4953]: I1203 00:08:11.781289 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:11 crc kubenswrapper[4953]: I1203 00:08:11.781311 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:11Z","lastTransitionTime":"2025-12-03T00:08:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:11 crc kubenswrapper[4953]: I1203 00:08:11.883637 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:11 crc kubenswrapper[4953]: I1203 00:08:11.883694 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:11 crc kubenswrapper[4953]: I1203 00:08:11.883710 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:11 crc kubenswrapper[4953]: I1203 00:08:11.883735 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:11 crc kubenswrapper[4953]: I1203 00:08:11.883754 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:11Z","lastTransitionTime":"2025-12-03T00:08:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:11 crc kubenswrapper[4953]: I1203 00:08:11.986832 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:11 crc kubenswrapper[4953]: I1203 00:08:11.986925 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:11 crc kubenswrapper[4953]: I1203 00:08:11.986938 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:11 crc kubenswrapper[4953]: I1203 00:08:11.986956 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:11 crc kubenswrapper[4953]: I1203 00:08:11.986967 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:11Z","lastTransitionTime":"2025-12-03T00:08:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:12 crc kubenswrapper[4953]: I1203 00:08:12.091482 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:12 crc kubenswrapper[4953]: I1203 00:08:12.091555 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:12 crc kubenswrapper[4953]: I1203 00:08:12.091579 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:12 crc kubenswrapper[4953]: I1203 00:08:12.091609 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:12 crc kubenswrapper[4953]: I1203 00:08:12.091631 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:12Z","lastTransitionTime":"2025-12-03T00:08:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:12 crc kubenswrapper[4953]: I1203 00:08:12.194918 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:12 crc kubenswrapper[4953]: I1203 00:08:12.194963 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:12 crc kubenswrapper[4953]: I1203 00:08:12.195010 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:12 crc kubenswrapper[4953]: I1203 00:08:12.195028 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:12 crc kubenswrapper[4953]: I1203 00:08:12.195039 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:12Z","lastTransitionTime":"2025-12-03T00:08:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:12 crc kubenswrapper[4953]: I1203 00:08:12.297708 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:12 crc kubenswrapper[4953]: I1203 00:08:12.297795 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:12 crc kubenswrapper[4953]: I1203 00:08:12.297819 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:12 crc kubenswrapper[4953]: I1203 00:08:12.297851 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:12 crc kubenswrapper[4953]: I1203 00:08:12.297872 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:12Z","lastTransitionTime":"2025-12-03T00:08:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:12 crc kubenswrapper[4953]: I1203 00:08:12.347743 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 00:08:12 crc kubenswrapper[4953]: E1203 00:08:12.347929 4953 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 00:08:12 crc kubenswrapper[4953]: I1203 00:08:12.348214 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 00:08:12 crc kubenswrapper[4953]: E1203 00:08:12.348373 4953 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 00:08:12 crc kubenswrapper[4953]: I1203 00:08:12.348510 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 00:08:12 crc kubenswrapper[4953]: E1203 00:08:12.348625 4953 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 00:08:12 crc kubenswrapper[4953]: I1203 00:08:12.349750 4953 scope.go:117] "RemoveContainer" containerID="dcb805f9ce285d3841e93d588fe7f3d42232382366d42b75bcfb129aefa91c1f" Dec 03 00:08:12 crc kubenswrapper[4953]: I1203 00:08:12.401623 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:12 crc kubenswrapper[4953]: I1203 00:08:12.401940 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:12 crc kubenswrapper[4953]: I1203 00:08:12.402222 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:12 crc kubenswrapper[4953]: I1203 00:08:12.402669 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:12 crc kubenswrapper[4953]: I1203 00:08:12.403133 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:12Z","lastTransitionTime":"2025-12-03T00:08:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:12 crc kubenswrapper[4953]: I1203 00:08:12.508102 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:12 crc kubenswrapper[4953]: I1203 00:08:12.508165 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:12 crc kubenswrapper[4953]: I1203 00:08:12.508182 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:12 crc kubenswrapper[4953]: I1203 00:08:12.508212 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:12 crc kubenswrapper[4953]: I1203 00:08:12.508229 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:12Z","lastTransitionTime":"2025-12-03T00:08:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:12 crc kubenswrapper[4953]: I1203 00:08:12.611879 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:12 crc kubenswrapper[4953]: I1203 00:08:12.611935 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:12 crc kubenswrapper[4953]: I1203 00:08:12.611945 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:12 crc kubenswrapper[4953]: I1203 00:08:12.611965 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:12 crc kubenswrapper[4953]: I1203 00:08:12.612013 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:12Z","lastTransitionTime":"2025-12-03T00:08:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:12 crc kubenswrapper[4953]: I1203 00:08:12.716018 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:12 crc kubenswrapper[4953]: I1203 00:08:12.716088 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:12 crc kubenswrapper[4953]: I1203 00:08:12.716108 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:12 crc kubenswrapper[4953]: I1203 00:08:12.716137 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:12 crc kubenswrapper[4953]: I1203 00:08:12.716154 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:12Z","lastTransitionTime":"2025-12-03T00:08:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:12 crc kubenswrapper[4953]: I1203 00:08:12.819509 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:12 crc kubenswrapper[4953]: I1203 00:08:12.819573 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:12 crc kubenswrapper[4953]: I1203 00:08:12.819600 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:12 crc kubenswrapper[4953]: I1203 00:08:12.819634 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:12 crc kubenswrapper[4953]: I1203 00:08:12.819657 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:12Z","lastTransitionTime":"2025-12-03T00:08:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:12 crc kubenswrapper[4953]: I1203 00:08:12.868785 4953 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-txffx_8816ebee-c735-4b8d-b1ec-65349707b347/ovnkube-controller/2.log" Dec 03 00:08:12 crc kubenswrapper[4953]: I1203 00:08:12.871376 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-txffx" event={"ID":"8816ebee-c735-4b8d-b1ec-65349707b347","Type":"ContainerStarted","Data":"0a4b625eda9b9daff8aa7fe3c5b94e417bcefe08fc6a65b4594f821cfdca7bfd"} Dec 03 00:08:12 crc kubenswrapper[4953]: I1203 00:08:12.872006 4953 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-txffx" Dec 03 00:08:12 crc kubenswrapper[4953]: I1203 00:08:12.891829 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:10Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:08:12Z is after 2025-08-24T17:21:41Z" Dec 03 00:08:12 crc kubenswrapper[4953]: I1203 00:08:12.914086 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-675qk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"43a4bad8-71f4-441a-8302-aeb526a986af\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:08:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:08:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e701c2ad05014683415ae2fb03b96f3acb38062ba7c5bfc4d0a9e34431695e36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5b5794fef6477aab5deed8f642e8cc50b37eb1fd0cc6f19c003f34ad31c1e3c0\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T00:08:02Z\\\",\\\"message\\\":\\\"2025-12-03T00:07:17+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_fa87bd16-684d-458f-b6b8-b6ada86b0d46\\\\n2025-12-03T00:07:17+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_fa87bd16-684d-458f-b6b8-b6ada86b0d46 to /host/opt/cni/bin/\\\\n2025-12-03T00:07:17Z [verbose] multus-daemon started\\\\n2025-12-03T00:07:17Z [verbose] Readiness Indicator file check\\\\n2025-12-03T00:08:02Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T00:07:16Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:08:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gdrw4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:07:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-675qk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:08:12Z is after 2025-08-24T17:21:41Z" Dec 03 00:08:12 crc kubenswrapper[4953]: I1203 00:08:12.922698 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:12 crc kubenswrapper[4953]: I1203 00:08:12.922735 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:12 crc kubenswrapper[4953]: I1203 00:08:12.922747 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:12 crc kubenswrapper[4953]: I1203 00:08:12.922767 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:12 crc kubenswrapper[4953]: I1203 00:08:12.922781 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:12Z","lastTransitionTime":"2025-12-03T00:08:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:12 crc kubenswrapper[4953]: I1203 00:08:12.927519 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f94f8c51-c665-4ca3-bf17-7a04ea5c7ade\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72c5cd7dc1c3a8adfedab6ba43a59d80b3efe65f2bb56be95e76cc86a409dcf3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2810ffc1b141fa7da811406053db9d72c708edf5f2bb1dd75e4d52ab35fd24c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c67b4277b20a9f1287f50ef82d0785c9d02543a8e7ed26e0413eae332b866db7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://258d78e1ad9d3dcda6976909ae3fe71aecead5b995fc66b3fd6e2722d80b5ab8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:06:51Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:08:12Z is after 2025-08-24T17:21:41Z" Dec 03 00:08:12 crc kubenswrapper[4953]: I1203 00:08:12.942723 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5652d0f3-3fd7-4af6-8941-9119e1177641\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2030aef5b7170991268fa5dc7e7ed8ce1fbe3a2765165a10c25251426575330f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ab6052f1abf17225155ce1bd86afa742e6e78f507b7deedd7161cd8fba9120b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://650b6c79cda2e1768ba1d8763db799e210ce2748893512068d25e60f821937ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d1d60162f7f6885d4f51b4ab44163cfae6d2d04355b1dd710e26b022450005df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d1d60162f7f6885d4f51b4ab44163cfae6d2d04355b1dd710e26b022450005df\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:06:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:06:51Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:08:12Z is after 2025-08-24T17:21:41Z" Dec 03 00:08:12 crc kubenswrapper[4953]: I1203 00:08:12.964223 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09657a9f448a4c92a759c73a0a892076d7eb46393e6fd565e5f73f34c71aac23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5af5eac4a58870f63497f99b503cfa5fb922278095e0332ec50376925ecf2109\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:08:12Z is after 2025-08-24T17:21:41Z" Dec 03 00:08:12 crc kubenswrapper[4953]: I1203 00:08:12.987742 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-wj6rz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"96aa67a0-15dd-420d-a1d5-ed239d7df1fb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01815cfd3f58c3f3f5f2a894c1790e0b69c5060c48dc0b1deff050b6ad7af916\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kf62m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b12bfb5bdf9fd4f2c50a335fc82df6ee0440eaf8741897da03aa2dae368cc6b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b12bfb5bdf9fd4f2c50a335fc82df6ee0440eaf8741897da03aa2dae368cc6b4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:07:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kf62m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83aaa5fa28110bc0ef05aff75cd1442bc9bbdd33b822ae2ef22d4e4ab83238bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://83aaa5fa28110bc0ef05aff75cd1442bc9bbdd33b822ae2ef22d4e4ab83238bc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:07:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:07:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kf62m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ab6213a33a19fedebce6d74a0ebbb3b3a2de5687504e21cc24093cc42ade635\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ab6213a33a19fedebce6d74a0ebbb3b3a2de5687504e21cc24093cc42ade635\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:07:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:07:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kf62m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://de480488843afa6b8dede1e1b42b5181dca1993e2f39dc1b461a18a57e3668e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://de480488843afa6b8dede1e1b42b5181dca1993e2f39dc1b461a18a57e3668e1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:07:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kf62m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://97be4ba692bc61078d4b0ba39fb0c12772548c6c9251ff3356ea702f151223cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://97be4ba692bc61078d4b0ba39fb0c12772548c6c9251ff3356ea702f151223cd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:07:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kf62m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1a1f009fcbd8af9217db1dc97f662719f8c1db232df0e2804e44c881abb1df2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b1a1f009fcbd8af9217db1dc97f662719f8c1db232df0e2804e44c881abb1df2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:07:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:07:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kf62m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:07:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-wj6rz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:08:12Z is after 2025-08-24T17:21:41Z" Dec 03 00:08:13 crc kubenswrapper[4953]: I1203 00:08:13.005294 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-kl5s9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d85e5c97-16ad-4bd7-9e73-b7f12678be97\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c21df91a9e4e3165686882b2eb559498224063dc633c3a63f958598108eca29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j24lb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:07:16Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-kl5s9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:08:13Z is after 2025-08-24T17:21:41Z" Dec 03 00:08:13 crc kubenswrapper[4953]: I1203 00:08:13.022118 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-xshk4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bd03a758-0ca9-47d8-86f7-18b2d610df43\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47gfh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47gfh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:07:31Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-xshk4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:08:13Z is after 2025-08-24T17:21:41Z" Dec 03 00:08:13 crc kubenswrapper[4953]: I1203 00:08:13.025052 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:13 crc kubenswrapper[4953]: I1203 00:08:13.025093 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:13 crc kubenswrapper[4953]: I1203 00:08:13.025106 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:13 crc kubenswrapper[4953]: I1203 00:08:13.025124 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:13 crc kubenswrapper[4953]: I1203 00:08:13.025136 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:13Z","lastTransitionTime":"2025-12-03T00:08:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:13 crc kubenswrapper[4953]: I1203 00:08:13.041099 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cf821840-5353-4620-8f73-99b56b371ab0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cae5e2994fcc24dc9b2ea73e79572615b3311edc42642b7f278fb84123989b23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://29e4b9d1b70a93fb133e0759ad148ae133766a66efeae3ff06a21dfae0e805ff\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dec43536772aa7c3f5a662fa7f3bd73b1cae2791ad454f85792287828d0ffeed\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://034e0bac6892d361700b8f33d6427b8a55e2bd3f89b14ea77a374c6be94d25e2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://761973393a93c59c23e69f41b8ac0d16abc35abd61ecbc6d51a42d94cd664233\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T00:07:09Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 00:07:03.713374 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 00:07:03.715052 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1528563048/tls.crt::/tmp/serving-cert-1528563048/tls.key\\\\\\\"\\\\nI1203 00:07:09.300052 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 00:07:09.304192 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 00:07:09.304255 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 00:07:09.304309 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 00:07:09.304323 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 00:07:09.313563 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 00:07:09.313597 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 00:07:09.313603 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 00:07:09.313609 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 00:07:09.313613 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 00:07:09.313618 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 00:07:09.313622 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 00:07:09.313709 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 00:07:09.327347 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T00:06:53Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c143d4232fb8bf7ef5999a664e431a00ab0db030f4d31a28d0bb21e00e38581\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:53Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72af6c27296eae2e275d455795c68f0d272869d63642b8353b17d466d0c901cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://72af6c27296eae2e275d455795c68f0d272869d63642b8353b17d466d0c901cf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:06:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:06:51Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:08:13Z is after 2025-08-24T17:21:41Z" Dec 03 00:08:13 crc kubenswrapper[4953]: I1203 00:08:13.062839 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:10Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:08:13Z is after 2025-08-24T17:21:41Z" Dec 03 00:08:13 crc kubenswrapper[4953]: I1203 00:08:13.088847 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce8b414e1fc22eb30decb6a6411b672e7f443055cb89ebb029fbf6ffc1d5dca7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:08:13Z is after 2025-08-24T17:21:41Z" Dec 03 00:08:13 crc kubenswrapper[4953]: I1203 00:08:13.104707 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-6dfbn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53ffb1e9-9a2e-4869-80d7-776f4aa7ef8b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ab4205c0d38ad6a0c32077eadb1fd5fd3a279d9c8b8a2ffbe693ec2c4018370c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ct26c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7df84405eda520ad5a22d35750edf07ef3c3977a756ad5739c54f91cba9da7b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ct26c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:07:16Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-6dfbn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:08:13Z is after 2025-08-24T17:21:41Z" Dec 03 00:08:13 crc kubenswrapper[4953]: I1203 00:08:13.126579 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-txffx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8816ebee-c735-4b8d-b1ec-65349707b347\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:16Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:16Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4b85224419a4075acb340e4b11095f2adb5c316e285afa4e19295c2181c0d24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0f6f645a58993f052a6f365bb808f62d99f7fea74886c5cc022b2d86112321a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a61b951a9cbab2c48c0f2df8351f97d7ff71bf3674f641fe5eac30155232b224\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd5f428de229749485267e81d34e31b708728b52474032e88460e54747f21e11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://00a2ea4a9bb870fce30e7fa845f787f184e7a19e42c95511655ea8c5f84a26d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc0a5d4cc9d614d8492626a1b114bdc52f510eb5965d842c02c083f08d4849a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a4b625eda9b9daff8aa7fe3c5b94e417bcefe08fc6a65b4594f821cfdca7bfd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dcb805f9ce285d3841e93d588fe7f3d42232382366d42b75bcfb129aefa91c1f\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T00:07:44Z\\\",\\\"message\\\":\\\"mers/externalversions/factory.go:140\\\\nI1203 00:07:44.043157 6594 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1203 00:07:44.043232 6594 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 00:07:44.043450 6594 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 00:07:44.043811 6594 reflector.go:311] Stopping reflector *v1.NetworkAttachmentDefinition (0s) from github.com/k8snetworkplumbingwg/network-attachment-definition-client/pkg/client/informers/externalversions/factory.go:117\\\\nI1203 00:07:44.044028 6594 reflector.go:311] Stopping reflector *v1.AdminPolicyBasedExternalRoute (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140\\\\nI1203 00:07:44.044197 6594 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI1203 00:07:44.044257 6594 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T00:07:43Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:08:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7deadc8e135ff1bc7c5a2f093e489bb5b4b6f166bc874c6acc2a841c5397f11e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af4c7764fca87144d01f3738ea010308744ccf50c54b7c471ccc213112de00d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af4c7764fca87144d01f3738ea010308744ccf50c54b7c471ccc213112de00d5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:07:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:07:16Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-txffx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:08:13Z is after 2025-08-24T17:21:41Z" Dec 03 00:08:13 crc kubenswrapper[4953]: I1203 00:08:13.127632 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:13 crc kubenswrapper[4953]: I1203 00:08:13.127666 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:13 crc kubenswrapper[4953]: I1203 00:08:13.127675 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:13 crc kubenswrapper[4953]: I1203 00:08:13.127695 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:13 crc kubenswrapper[4953]: I1203 00:08:13.127705 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:13Z","lastTransitionTime":"2025-12-03T00:08:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:13 crc kubenswrapper[4953]: I1203 00:08:13.140554 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5cmlc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7e11a1d3-ecee-4bf0-87bf-085fab97b73d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d42390b9c53efc63add8ddcebf03c976f5e5448eaa2c34e076b2dc90dbac60fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6x2sc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:07:18Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5cmlc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:08:13Z is after 2025-08-24T17:21:41Z" Dec 03 00:08:13 crc kubenswrapper[4953]: I1203 00:08:13.150715 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-chmnz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a12fe6d-d25a-4b08-90da-cc9e510980a7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://711ab839cb9c0a90554dadbd45d810fc78aac09739340482fef8de8743b77283\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c982g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f802d5628a8cc32618044b9383211d5a492c56060d427f953b7793a1064029e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c982g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:07:29Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-chmnz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:08:13Z is after 2025-08-24T17:21:41Z" Dec 03 00:08:13 crc kubenswrapper[4953]: I1203 00:08:13.160826 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7cf23d2d-21d4-493d-bdc7-d088aa876304\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4c95eb1f9b0c797376a1b557415b3f956890c971dc77c5257f238b4f32ff9b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ad2a6f5a3296c6a4f53f9da696e3e3f4d6c919f44b380c9b4114021b8ad74b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ad2a6f5a3296c6a4f53f9da696e3e3f4d6c919f44b380c9b4114021b8ad74b6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:06:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:06:51Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:08:13Z is after 2025-08-24T17:21:41Z" Dec 03 00:08:13 crc kubenswrapper[4953]: I1203 00:08:13.172685 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:08:13Z is after 2025-08-24T17:21:41Z" Dec 03 00:08:13 crc kubenswrapper[4953]: I1203 00:08:13.188863 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2ed9c7fa96c8951fbc1728fe7a91fefdd999077285621df9c1abedf9ae0c449\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:08:13Z is after 2025-08-24T17:21:41Z" Dec 03 00:08:13 crc kubenswrapper[4953]: I1203 00:08:13.230072 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:13 crc kubenswrapper[4953]: I1203 00:08:13.230126 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:13 crc kubenswrapper[4953]: I1203 00:08:13.230139 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:13 crc kubenswrapper[4953]: I1203 00:08:13.230156 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:13 crc kubenswrapper[4953]: I1203 00:08:13.230168 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:13Z","lastTransitionTime":"2025-12-03T00:08:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:13 crc kubenswrapper[4953]: I1203 00:08:13.333126 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:13 crc kubenswrapper[4953]: I1203 00:08:13.333174 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:13 crc kubenswrapper[4953]: I1203 00:08:13.333186 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:13 crc kubenswrapper[4953]: I1203 00:08:13.333205 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:13 crc kubenswrapper[4953]: I1203 00:08:13.333220 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:13Z","lastTransitionTime":"2025-12-03T00:08:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:13 crc kubenswrapper[4953]: I1203 00:08:13.347564 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xshk4" Dec 03 00:08:13 crc kubenswrapper[4953]: E1203 00:08:13.347713 4953 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xshk4" podUID="bd03a758-0ca9-47d8-86f7-18b2d610df43" Dec 03 00:08:13 crc kubenswrapper[4953]: I1203 00:08:13.435722 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:13 crc kubenswrapper[4953]: I1203 00:08:13.435782 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:13 crc kubenswrapper[4953]: I1203 00:08:13.435798 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:13 crc kubenswrapper[4953]: I1203 00:08:13.435819 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:13 crc kubenswrapper[4953]: I1203 00:08:13.435832 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:13Z","lastTransitionTime":"2025-12-03T00:08:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:13 crc kubenswrapper[4953]: I1203 00:08:13.538494 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:13 crc kubenswrapper[4953]: I1203 00:08:13.538543 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:13 crc kubenswrapper[4953]: I1203 00:08:13.538553 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:13 crc kubenswrapper[4953]: I1203 00:08:13.538569 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:13 crc kubenswrapper[4953]: I1203 00:08:13.538580 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:13Z","lastTransitionTime":"2025-12-03T00:08:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:13 crc kubenswrapper[4953]: I1203 00:08:13.641146 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:13 crc kubenswrapper[4953]: I1203 00:08:13.641179 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:13 crc kubenswrapper[4953]: I1203 00:08:13.641189 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:13 crc kubenswrapper[4953]: I1203 00:08:13.641204 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:13 crc kubenswrapper[4953]: I1203 00:08:13.641214 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:13Z","lastTransitionTime":"2025-12-03T00:08:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:13 crc kubenswrapper[4953]: I1203 00:08:13.744231 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:13 crc kubenswrapper[4953]: I1203 00:08:13.744296 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:13 crc kubenswrapper[4953]: I1203 00:08:13.744318 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:13 crc kubenswrapper[4953]: I1203 00:08:13.744349 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:13 crc kubenswrapper[4953]: I1203 00:08:13.744371 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:13Z","lastTransitionTime":"2025-12-03T00:08:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:13 crc kubenswrapper[4953]: I1203 00:08:13.847855 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:13 crc kubenswrapper[4953]: I1203 00:08:13.847903 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:13 crc kubenswrapper[4953]: I1203 00:08:13.847918 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:13 crc kubenswrapper[4953]: I1203 00:08:13.847937 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:13 crc kubenswrapper[4953]: I1203 00:08:13.847948 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:13Z","lastTransitionTime":"2025-12-03T00:08:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:13 crc kubenswrapper[4953]: I1203 00:08:13.878056 4953 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-txffx_8816ebee-c735-4b8d-b1ec-65349707b347/ovnkube-controller/3.log" Dec 03 00:08:13 crc kubenswrapper[4953]: I1203 00:08:13.878926 4953 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-txffx_8816ebee-c735-4b8d-b1ec-65349707b347/ovnkube-controller/2.log" Dec 03 00:08:13 crc kubenswrapper[4953]: I1203 00:08:13.883150 4953 generic.go:334] "Generic (PLEG): container finished" podID="8816ebee-c735-4b8d-b1ec-65349707b347" containerID="0a4b625eda9b9daff8aa7fe3c5b94e417bcefe08fc6a65b4594f821cfdca7bfd" exitCode=1 Dec 03 00:08:13 crc kubenswrapper[4953]: I1203 00:08:13.883229 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-txffx" event={"ID":"8816ebee-c735-4b8d-b1ec-65349707b347","Type":"ContainerDied","Data":"0a4b625eda9b9daff8aa7fe3c5b94e417bcefe08fc6a65b4594f821cfdca7bfd"} Dec 03 00:08:13 crc kubenswrapper[4953]: I1203 00:08:13.883284 4953 scope.go:117] "RemoveContainer" containerID="dcb805f9ce285d3841e93d588fe7f3d42232382366d42b75bcfb129aefa91c1f" Dec 03 00:08:13 crc kubenswrapper[4953]: I1203 00:08:13.885519 4953 scope.go:117] "RemoveContainer" containerID="0a4b625eda9b9daff8aa7fe3c5b94e417bcefe08fc6a65b4594f821cfdca7bfd" Dec 03 00:08:13 crc kubenswrapper[4953]: E1203 00:08:13.885833 4953 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-txffx_openshift-ovn-kubernetes(8816ebee-c735-4b8d-b1ec-65349707b347)\"" pod="openshift-ovn-kubernetes/ovnkube-node-txffx" podUID="8816ebee-c735-4b8d-b1ec-65349707b347" Dec 03 00:08:13 crc kubenswrapper[4953]: I1203 00:08:13.908355 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f94f8c51-c665-4ca3-bf17-7a04ea5c7ade\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72c5cd7dc1c3a8adfedab6ba43a59d80b3efe65f2bb56be95e76cc86a409dcf3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2810ffc1b141fa7da811406053db9d72c708edf5f2bb1dd75e4d52ab35fd24c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c67b4277b20a9f1287f50ef82d0785c9d02543a8e7ed26e0413eae332b866db7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://258d78e1ad9d3dcda6976909ae3fe71aecead5b995fc66b3fd6e2722d80b5ab8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:06:51Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:08:13Z is after 2025-08-24T17:21:41Z" Dec 03 00:08:13 crc kubenswrapper[4953]: I1203 00:08:13.925730 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5652d0f3-3fd7-4af6-8941-9119e1177641\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2030aef5b7170991268fa5dc7e7ed8ce1fbe3a2765165a10c25251426575330f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ab6052f1abf17225155ce1bd86afa742e6e78f507b7deedd7161cd8fba9120b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://650b6c79cda2e1768ba1d8763db799e210ce2748893512068d25e60f821937ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d1d60162f7f6885d4f51b4ab44163cfae6d2d04355b1dd710e26b022450005df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d1d60162f7f6885d4f51b4ab44163cfae6d2d04355b1dd710e26b022450005df\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:06:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:06:51Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:08:13Z is after 2025-08-24T17:21:41Z" Dec 03 00:08:13 crc kubenswrapper[4953]: I1203 00:08:13.948255 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09657a9f448a4c92a759c73a0a892076d7eb46393e6fd565e5f73f34c71aac23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5af5eac4a58870f63497f99b503cfa5fb922278095e0332ec50376925ecf2109\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:08:13Z is after 2025-08-24T17:21:41Z" Dec 03 00:08:13 crc kubenswrapper[4953]: I1203 00:08:13.950757 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:13 crc kubenswrapper[4953]: I1203 00:08:13.950797 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:13 crc kubenswrapper[4953]: I1203 00:08:13.950807 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:13 crc kubenswrapper[4953]: I1203 00:08:13.950823 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:13 crc kubenswrapper[4953]: I1203 00:08:13.950834 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:13Z","lastTransitionTime":"2025-12-03T00:08:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:13 crc kubenswrapper[4953]: I1203 00:08:13.968912 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-wj6rz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"96aa67a0-15dd-420d-a1d5-ed239d7df1fb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01815cfd3f58c3f3f5f2a894c1790e0b69c5060c48dc0b1deff050b6ad7af916\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kf62m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b12bfb5bdf9fd4f2c50a335fc82df6ee0440eaf8741897da03aa2dae368cc6b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b12bfb5bdf9fd4f2c50a335fc82df6ee0440eaf8741897da03aa2dae368cc6b4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:07:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kf62m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83aaa5fa28110bc0ef05aff75cd1442bc9bbdd33b822ae2ef22d4e4ab83238bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://83aaa5fa28110bc0ef05aff75cd1442bc9bbdd33b822ae2ef22d4e4ab83238bc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:07:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:07:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kf62m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ab6213a33a19fedebce6d74a0ebbb3b3a2de5687504e21cc24093cc42ade635\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ab6213a33a19fedebce6d74a0ebbb3b3a2de5687504e21cc24093cc42ade635\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:07:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:07:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kf62m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://de480488843afa6b8dede1e1b42b5181dca1993e2f39dc1b461a18a57e3668e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://de480488843afa6b8dede1e1b42b5181dca1993e2f39dc1b461a18a57e3668e1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:07:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kf62m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://97be4ba692bc61078d4b0ba39fb0c12772548c6c9251ff3356ea702f151223cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://97be4ba692bc61078d4b0ba39fb0c12772548c6c9251ff3356ea702f151223cd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:07:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kf62m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1a1f009fcbd8af9217db1dc97f662719f8c1db232df0e2804e44c881abb1df2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b1a1f009fcbd8af9217db1dc97f662719f8c1db232df0e2804e44c881abb1df2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:07:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:07:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kf62m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:07:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-wj6rz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:08:13Z is after 2025-08-24T17:21:41Z" Dec 03 00:08:13 crc kubenswrapper[4953]: I1203 00:08:13.978244 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-kl5s9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d85e5c97-16ad-4bd7-9e73-b7f12678be97\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c21df91a9e4e3165686882b2eb559498224063dc633c3a63f958598108eca29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j24lb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:07:16Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-kl5s9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:08:13Z is after 2025-08-24T17:21:41Z" Dec 03 00:08:13 crc kubenswrapper[4953]: I1203 00:08:13.986798 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-chmnz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a12fe6d-d25a-4b08-90da-cc9e510980a7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://711ab839cb9c0a90554dadbd45d810fc78aac09739340482fef8de8743b77283\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c982g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f802d5628a8cc32618044b9383211d5a492c56060d427f953b7793a1064029e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c982g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:07:29Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-chmnz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:08:13Z is after 2025-08-24T17:21:41Z" Dec 03 00:08:13 crc kubenswrapper[4953]: I1203 00:08:13.996958 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-xshk4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bd03a758-0ca9-47d8-86f7-18b2d610df43\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47gfh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47gfh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:07:31Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-xshk4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:08:13Z is after 2025-08-24T17:21:41Z" Dec 03 00:08:14 crc kubenswrapper[4953]: I1203 00:08:14.008625 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cf821840-5353-4620-8f73-99b56b371ab0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cae5e2994fcc24dc9b2ea73e79572615b3311edc42642b7f278fb84123989b23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://29e4b9d1b70a93fb133e0759ad148ae133766a66efeae3ff06a21dfae0e805ff\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dec43536772aa7c3f5a662fa7f3bd73b1cae2791ad454f85792287828d0ffeed\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://034e0bac6892d361700b8f33d6427b8a55e2bd3f89b14ea77a374c6be94d25e2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://761973393a93c59c23e69f41b8ac0d16abc35abd61ecbc6d51a42d94cd664233\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T00:07:09Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 00:07:03.713374 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 00:07:03.715052 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1528563048/tls.crt::/tmp/serving-cert-1528563048/tls.key\\\\\\\"\\\\nI1203 00:07:09.300052 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 00:07:09.304192 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 00:07:09.304255 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 00:07:09.304309 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 00:07:09.304323 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 00:07:09.313563 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 00:07:09.313597 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 00:07:09.313603 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 00:07:09.313609 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 00:07:09.313613 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 00:07:09.313618 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 00:07:09.313622 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 00:07:09.313709 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 00:07:09.327347 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T00:06:53Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c143d4232fb8bf7ef5999a664e431a00ab0db030f4d31a28d0bb21e00e38581\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:53Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72af6c27296eae2e275d455795c68f0d272869d63642b8353b17d466d0c901cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://72af6c27296eae2e275d455795c68f0d272869d63642b8353b17d466d0c901cf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:06:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:06:51Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:08:14Z is after 2025-08-24T17:21:41Z" Dec 03 00:08:14 crc kubenswrapper[4953]: I1203 00:08:14.019586 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:10Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:08:14Z is after 2025-08-24T17:21:41Z" Dec 03 00:08:14 crc kubenswrapper[4953]: I1203 00:08:14.034689 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce8b414e1fc22eb30decb6a6411b672e7f443055cb89ebb029fbf6ffc1d5dca7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:08:14Z is after 2025-08-24T17:21:41Z" Dec 03 00:08:14 crc kubenswrapper[4953]: I1203 00:08:14.045554 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-6dfbn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53ffb1e9-9a2e-4869-80d7-776f4aa7ef8b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ab4205c0d38ad6a0c32077eadb1fd5fd3a279d9c8b8a2ffbe693ec2c4018370c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ct26c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7df84405eda520ad5a22d35750edf07ef3c3977a756ad5739c54f91cba9da7b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ct26c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:07:16Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-6dfbn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:08:14Z is after 2025-08-24T17:21:41Z" Dec 03 00:08:14 crc kubenswrapper[4953]: I1203 00:08:14.053052 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:14 crc kubenswrapper[4953]: I1203 00:08:14.053120 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:14 crc kubenswrapper[4953]: I1203 00:08:14.053147 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:14 crc kubenswrapper[4953]: I1203 00:08:14.053172 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:14 crc kubenswrapper[4953]: I1203 00:08:14.053195 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:14Z","lastTransitionTime":"2025-12-03T00:08:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:14 crc kubenswrapper[4953]: I1203 00:08:14.064587 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-txffx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8816ebee-c735-4b8d-b1ec-65349707b347\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:16Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:16Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4b85224419a4075acb340e4b11095f2adb5c316e285afa4e19295c2181c0d24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0f6f645a58993f052a6f365bb808f62d99f7fea74886c5cc022b2d86112321a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a61b951a9cbab2c48c0f2df8351f97d7ff71bf3674f641fe5eac30155232b224\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd5f428de229749485267e81d34e31b708728b52474032e88460e54747f21e11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://00a2ea4a9bb870fce30e7fa845f787f184e7a19e42c95511655ea8c5f84a26d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc0a5d4cc9d614d8492626a1b114bdc52f510eb5965d842c02c083f08d4849a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a4b625eda9b9daff8aa7fe3c5b94e417bcefe08fc6a65b4594f821cfdca7bfd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dcb805f9ce285d3841e93d588fe7f3d42232382366d42b75bcfb129aefa91c1f\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T00:07:44Z\\\",\\\"message\\\":\\\"mers/externalversions/factory.go:140\\\\nI1203 00:07:44.043157 6594 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1203 00:07:44.043232 6594 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 00:07:44.043450 6594 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 00:07:44.043811 6594 reflector.go:311] Stopping reflector *v1.NetworkAttachmentDefinition (0s) from github.com/k8snetworkplumbingwg/network-attachment-definition-client/pkg/client/informers/externalversions/factory.go:117\\\\nI1203 00:07:44.044028 6594 reflector.go:311] Stopping reflector *v1.AdminPolicyBasedExternalRoute (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140\\\\nI1203 00:07:44.044197 6594 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI1203 00:07:44.044257 6594 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T00:07:43Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0a4b625eda9b9daff8aa7fe3c5b94e417bcefe08fc6a65b4594f821cfdca7bfd\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T00:08:13Z\\\",\\\"message\\\":\\\"il\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1203 00:08:13.521854 6989 transact.go:42] Configuring OVN: [{Op:update Table:Logical_Router_Static_Route Row:map[ip_prefix:10.217.0.0/22 nexthop:100.64.0.2 policy:{GoSet:[src-ip]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {8944024f-deb7-4076-afb3-4b50a2ff4b4b}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:mutate Table:Logical_Router Row:map[] Rows:[] Columns:[] Mutations:[{Column:static_routes Mutator:insert Value:{GoSet:[{GoUUID:8944024f-deb7-4076-afb3-4b50a2ff4b4b}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {f6d604c1-9711-4e25-be6c-79ec28bbad1b}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1203 00:08:13.522171 6989 obj_retry.go:551] Creating *factory.egressNode crc took: 1.771964ms\\\\nI1203 00:08:13.522207 6989 factory.go:1336] Added *v1.Node event handler 7\\\\nI1203 00:08:13.522242 6989 factory.go:1336] Added *v1.EgressIP event handler 8\\\\nI1203 00:08:13.522575 6989 factory.go:1336] Added *v1.EgressFirewall event handler 9\\\\nI1203 00:08:13.522717 6989 controller.go:132] Adding controller ef_node_controller event handlers\\\\nI1203 00:08:13.522757 6989 ovnkube.go:599] Stopped ovnkube\\\\nI1203 00:08:13.522794 6989 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1203 00:08:13.522871 6989 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T00:08:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7deadc8e135ff1bc7c5a2f093e489bb5b4b6f166bc874c6acc2a841c5397f11e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af4c7764fca87144d01f3738ea010308744ccf50c54b7c471ccc213112de00d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af4c7764fca87144d01f3738ea010308744ccf50c54b7c471ccc213112de00d5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:07:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:07:16Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-txffx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:08:14Z is after 2025-08-24T17:21:41Z" Dec 03 00:08:14 crc kubenswrapper[4953]: I1203 00:08:14.078259 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5cmlc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7e11a1d3-ecee-4bf0-87bf-085fab97b73d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d42390b9c53efc63add8ddcebf03c976f5e5448eaa2c34e076b2dc90dbac60fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6x2sc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:07:18Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5cmlc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:08:14Z is after 2025-08-24T17:21:41Z" Dec 03 00:08:14 crc kubenswrapper[4953]: I1203 00:08:14.092108 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7cf23d2d-21d4-493d-bdc7-d088aa876304\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4c95eb1f9b0c797376a1b557415b3f956890c971dc77c5257f238b4f32ff9b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ad2a6f5a3296c6a4f53f9da696e3e3f4d6c919f44b380c9b4114021b8ad74b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ad2a6f5a3296c6a4f53f9da696e3e3f4d6c919f44b380c9b4114021b8ad74b6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:06:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:06:51Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:08:14Z is after 2025-08-24T17:21:41Z" Dec 03 00:08:14 crc kubenswrapper[4953]: I1203 00:08:14.105424 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:08:14Z is after 2025-08-24T17:21:41Z" Dec 03 00:08:14 crc kubenswrapper[4953]: I1203 00:08:14.122649 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2ed9c7fa96c8951fbc1728fe7a91fefdd999077285621df9c1abedf9ae0c449\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:08:14Z is after 2025-08-24T17:21:41Z" Dec 03 00:08:14 crc kubenswrapper[4953]: I1203 00:08:14.136094 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:10Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:08:14Z is after 2025-08-24T17:21:41Z" Dec 03 00:08:14 crc kubenswrapper[4953]: I1203 00:08:14.152028 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-675qk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"43a4bad8-71f4-441a-8302-aeb526a986af\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:08:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:08:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e701c2ad05014683415ae2fb03b96f3acb38062ba7c5bfc4d0a9e34431695e36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5b5794fef6477aab5deed8f642e8cc50b37eb1fd0cc6f19c003f34ad31c1e3c0\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T00:08:02Z\\\",\\\"message\\\":\\\"2025-12-03T00:07:17+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_fa87bd16-684d-458f-b6b8-b6ada86b0d46\\\\n2025-12-03T00:07:17+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_fa87bd16-684d-458f-b6b8-b6ada86b0d46 to /host/opt/cni/bin/\\\\n2025-12-03T00:07:17Z [verbose] multus-daemon started\\\\n2025-12-03T00:07:17Z [verbose] Readiness Indicator file check\\\\n2025-12-03T00:08:02Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T00:07:16Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:08:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gdrw4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:07:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-675qk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:08:14Z is after 2025-08-24T17:21:41Z" Dec 03 00:08:14 crc kubenswrapper[4953]: I1203 00:08:14.155475 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:14 crc kubenswrapper[4953]: I1203 00:08:14.155539 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:14 crc kubenswrapper[4953]: I1203 00:08:14.155555 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:14 crc kubenswrapper[4953]: I1203 00:08:14.155580 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:14 crc kubenswrapper[4953]: I1203 00:08:14.155597 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:14Z","lastTransitionTime":"2025-12-03T00:08:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:14 crc kubenswrapper[4953]: I1203 00:08:14.258707 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:14 crc kubenswrapper[4953]: I1203 00:08:14.258756 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:14 crc kubenswrapper[4953]: I1203 00:08:14.258767 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:14 crc kubenswrapper[4953]: I1203 00:08:14.258782 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:14 crc kubenswrapper[4953]: I1203 00:08:14.258794 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:14Z","lastTransitionTime":"2025-12-03T00:08:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:14 crc kubenswrapper[4953]: I1203 00:08:14.267416 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 00:08:14 crc kubenswrapper[4953]: I1203 00:08:14.267513 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 00:08:14 crc kubenswrapper[4953]: I1203 00:08:14.267542 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 00:08:14 crc kubenswrapper[4953]: E1203 00:08:14.267573 4953 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 00:09:18.267557649 +0000 UTC m=+147.251838432 (durationBeforeRetry 1m4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 00:08:14 crc kubenswrapper[4953]: I1203 00:08:14.267595 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 00:08:14 crc kubenswrapper[4953]: I1203 00:08:14.267622 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 00:08:14 crc kubenswrapper[4953]: E1203 00:08:14.267664 4953 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 03 00:08:14 crc kubenswrapper[4953]: E1203 00:08:14.267680 4953 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 03 00:08:14 crc kubenswrapper[4953]: E1203 00:08:14.267693 4953 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 00:08:14 crc kubenswrapper[4953]: E1203 00:08:14.267711 4953 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 03 00:08:14 crc kubenswrapper[4953]: E1203 00:08:14.267670 4953 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 03 00:08:14 crc kubenswrapper[4953]: E1203 00:08:14.267733 4953 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-03 00:09:18.267721113 +0000 UTC m=+147.252001916 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 00:08:14 crc kubenswrapper[4953]: E1203 00:08:14.267739 4953 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 03 00:08:14 crc kubenswrapper[4953]: E1203 00:08:14.267779 4953 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 03 00:08:14 crc kubenswrapper[4953]: E1203 00:08:14.267797 4953 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 00:08:14 crc kubenswrapper[4953]: E1203 00:08:14.267758 4953 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-03 00:09:18.267751054 +0000 UTC m=+147.252031837 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 03 00:08:14 crc kubenswrapper[4953]: E1203 00:08:14.267854 4953 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-03 00:09:18.267847436 +0000 UTC m=+147.252128219 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 03 00:08:14 crc kubenswrapper[4953]: E1203 00:08:14.267867 4953 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-03 00:09:18.267861466 +0000 UTC m=+147.252142249 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 00:08:14 crc kubenswrapper[4953]: I1203 00:08:14.348121 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 00:08:14 crc kubenswrapper[4953]: I1203 00:08:14.348128 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 00:08:14 crc kubenswrapper[4953]: I1203 00:08:14.348310 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 00:08:14 crc kubenswrapper[4953]: E1203 00:08:14.348406 4953 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 00:08:14 crc kubenswrapper[4953]: E1203 00:08:14.348492 4953 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 00:08:14 crc kubenswrapper[4953]: E1203 00:08:14.348600 4953 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 00:08:14 crc kubenswrapper[4953]: I1203 00:08:14.361385 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:14 crc kubenswrapper[4953]: I1203 00:08:14.361428 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:14 crc kubenswrapper[4953]: I1203 00:08:14.361444 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:14 crc kubenswrapper[4953]: I1203 00:08:14.361462 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:14 crc kubenswrapper[4953]: I1203 00:08:14.361474 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:14Z","lastTransitionTime":"2025-12-03T00:08:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:14 crc kubenswrapper[4953]: I1203 00:08:14.422365 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:14 crc kubenswrapper[4953]: I1203 00:08:14.422404 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:14 crc kubenswrapper[4953]: I1203 00:08:14.422413 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:14 crc kubenswrapper[4953]: I1203 00:08:14.422428 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:14 crc kubenswrapper[4953]: I1203 00:08:14.422439 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:14Z","lastTransitionTime":"2025-12-03T00:08:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:14 crc kubenswrapper[4953]: E1203 00:08:14.441799 4953 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:08:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:08:14Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:08:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:08:14Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:08:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:08:14Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:08:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:08:14Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"eb4ae5a6-0f6f-49ef-a027-1cdee9d3c536\\\",\\\"systemUUID\\\":\\\"4c1f3d44-3d3d-425d-993f-471223206a89\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:08:14Z is after 2025-08-24T17:21:41Z" Dec 03 00:08:14 crc kubenswrapper[4953]: I1203 00:08:14.447421 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:14 crc kubenswrapper[4953]: I1203 00:08:14.447486 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:14 crc kubenswrapper[4953]: I1203 00:08:14.447509 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:14 crc kubenswrapper[4953]: I1203 00:08:14.447540 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:14 crc kubenswrapper[4953]: I1203 00:08:14.447562 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:14Z","lastTransitionTime":"2025-12-03T00:08:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:14 crc kubenswrapper[4953]: E1203 00:08:14.460761 4953 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:08:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:08:14Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:08:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:08:14Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:08:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:08:14Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:08:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:08:14Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"eb4ae5a6-0f6f-49ef-a027-1cdee9d3c536\\\",\\\"systemUUID\\\":\\\"4c1f3d44-3d3d-425d-993f-471223206a89\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:08:14Z is after 2025-08-24T17:21:41Z" Dec 03 00:08:14 crc kubenswrapper[4953]: I1203 00:08:14.464464 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:14 crc kubenswrapper[4953]: I1203 00:08:14.464516 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:14 crc kubenswrapper[4953]: I1203 00:08:14.464529 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:14 crc kubenswrapper[4953]: I1203 00:08:14.464550 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:14 crc kubenswrapper[4953]: I1203 00:08:14.464561 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:14Z","lastTransitionTime":"2025-12-03T00:08:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:14 crc kubenswrapper[4953]: E1203 00:08:14.477509 4953 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:08:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:08:14Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:08:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:08:14Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:08:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:08:14Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:08:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:08:14Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"eb4ae5a6-0f6f-49ef-a027-1cdee9d3c536\\\",\\\"systemUUID\\\":\\\"4c1f3d44-3d3d-425d-993f-471223206a89\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:08:14Z is after 2025-08-24T17:21:41Z" Dec 03 00:08:14 crc kubenswrapper[4953]: I1203 00:08:14.481362 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:14 crc kubenswrapper[4953]: I1203 00:08:14.481401 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:14 crc kubenswrapper[4953]: I1203 00:08:14.481412 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:14 crc kubenswrapper[4953]: I1203 00:08:14.481425 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:14 crc kubenswrapper[4953]: I1203 00:08:14.481435 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:14Z","lastTransitionTime":"2025-12-03T00:08:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:14 crc kubenswrapper[4953]: E1203 00:08:14.494346 4953 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:08:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:08:14Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:08:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:08:14Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:08:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:08:14Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:08:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:08:14Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"eb4ae5a6-0f6f-49ef-a027-1cdee9d3c536\\\",\\\"systemUUID\\\":\\\"4c1f3d44-3d3d-425d-993f-471223206a89\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:08:14Z is after 2025-08-24T17:21:41Z" Dec 03 00:08:14 crc kubenswrapper[4953]: I1203 00:08:14.497540 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:14 crc kubenswrapper[4953]: I1203 00:08:14.497609 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:14 crc kubenswrapper[4953]: I1203 00:08:14.497631 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:14 crc kubenswrapper[4953]: I1203 00:08:14.497659 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:14 crc kubenswrapper[4953]: I1203 00:08:14.497680 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:14Z","lastTransitionTime":"2025-12-03T00:08:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:14 crc kubenswrapper[4953]: E1203 00:08:14.514943 4953 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:08:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:08:14Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:08:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:08:14Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:08:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:08:14Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:08:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:08:14Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"eb4ae5a6-0f6f-49ef-a027-1cdee9d3c536\\\",\\\"systemUUID\\\":\\\"4c1f3d44-3d3d-425d-993f-471223206a89\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:08:14Z is after 2025-08-24T17:21:41Z" Dec 03 00:08:14 crc kubenswrapper[4953]: E1203 00:08:14.515085 4953 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 03 00:08:14 crc kubenswrapper[4953]: I1203 00:08:14.516732 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:14 crc kubenswrapper[4953]: I1203 00:08:14.516757 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:14 crc kubenswrapper[4953]: I1203 00:08:14.516766 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:14 crc kubenswrapper[4953]: I1203 00:08:14.516780 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:14 crc kubenswrapper[4953]: I1203 00:08:14.516790 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:14Z","lastTransitionTime":"2025-12-03T00:08:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:14 crc kubenswrapper[4953]: I1203 00:08:14.619425 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:14 crc kubenswrapper[4953]: I1203 00:08:14.619462 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:14 crc kubenswrapper[4953]: I1203 00:08:14.619496 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:14 crc kubenswrapper[4953]: I1203 00:08:14.619512 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:14 crc kubenswrapper[4953]: I1203 00:08:14.619523 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:14Z","lastTransitionTime":"2025-12-03T00:08:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:14 crc kubenswrapper[4953]: I1203 00:08:14.721808 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:14 crc kubenswrapper[4953]: I1203 00:08:14.722065 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:14 crc kubenswrapper[4953]: I1203 00:08:14.722197 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:14 crc kubenswrapper[4953]: I1203 00:08:14.722304 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:14 crc kubenswrapper[4953]: I1203 00:08:14.722396 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:14Z","lastTransitionTime":"2025-12-03T00:08:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:14 crc kubenswrapper[4953]: I1203 00:08:14.825228 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:14 crc kubenswrapper[4953]: I1203 00:08:14.825289 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:14 crc kubenswrapper[4953]: I1203 00:08:14.825306 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:14 crc kubenswrapper[4953]: I1203 00:08:14.825332 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:14 crc kubenswrapper[4953]: I1203 00:08:14.825352 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:14Z","lastTransitionTime":"2025-12-03T00:08:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:14 crc kubenswrapper[4953]: I1203 00:08:14.890860 4953 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-txffx_8816ebee-c735-4b8d-b1ec-65349707b347/ovnkube-controller/3.log" Dec 03 00:08:14 crc kubenswrapper[4953]: I1203 00:08:14.898287 4953 scope.go:117] "RemoveContainer" containerID="0a4b625eda9b9daff8aa7fe3c5b94e417bcefe08fc6a65b4594f821cfdca7bfd" Dec 03 00:08:14 crc kubenswrapper[4953]: E1203 00:08:14.898590 4953 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-txffx_openshift-ovn-kubernetes(8816ebee-c735-4b8d-b1ec-65349707b347)\"" pod="openshift-ovn-kubernetes/ovnkube-node-txffx" podUID="8816ebee-c735-4b8d-b1ec-65349707b347" Dec 03 00:08:14 crc kubenswrapper[4953]: I1203 00:08:14.916739 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7cf23d2d-21d4-493d-bdc7-d088aa876304\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4c95eb1f9b0c797376a1b557415b3f956890c971dc77c5257f238b4f32ff9b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ad2a6f5a3296c6a4f53f9da696e3e3f4d6c919f44b380c9b4114021b8ad74b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ad2a6f5a3296c6a4f53f9da696e3e3f4d6c919f44b380c9b4114021b8ad74b6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:06:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:06:51Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:08:14Z is after 2025-08-24T17:21:41Z" Dec 03 00:08:14 crc kubenswrapper[4953]: I1203 00:08:14.929190 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:14 crc kubenswrapper[4953]: I1203 00:08:14.929492 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:14 crc kubenswrapper[4953]: I1203 00:08:14.929625 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:14 crc kubenswrapper[4953]: I1203 00:08:14.929758 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:14 crc kubenswrapper[4953]: I1203 00:08:14.929931 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:14Z","lastTransitionTime":"2025-12-03T00:08:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:14 crc kubenswrapper[4953]: I1203 00:08:14.937512 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:08:14Z is after 2025-08-24T17:21:41Z" Dec 03 00:08:14 crc kubenswrapper[4953]: I1203 00:08:14.956691 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2ed9c7fa96c8951fbc1728fe7a91fefdd999077285621df9c1abedf9ae0c449\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:08:14Z is after 2025-08-24T17:21:41Z" Dec 03 00:08:14 crc kubenswrapper[4953]: I1203 00:08:14.977060 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:10Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:08:14Z is after 2025-08-24T17:21:41Z" Dec 03 00:08:14 crc kubenswrapper[4953]: I1203 00:08:14.993291 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-675qk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"43a4bad8-71f4-441a-8302-aeb526a986af\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:08:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:08:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e701c2ad05014683415ae2fb03b96f3acb38062ba7c5bfc4d0a9e34431695e36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5b5794fef6477aab5deed8f642e8cc50b37eb1fd0cc6f19c003f34ad31c1e3c0\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T00:08:02Z\\\",\\\"message\\\":\\\"2025-12-03T00:07:17+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_fa87bd16-684d-458f-b6b8-b6ada86b0d46\\\\n2025-12-03T00:07:17+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_fa87bd16-684d-458f-b6b8-b6ada86b0d46 to /host/opt/cni/bin/\\\\n2025-12-03T00:07:17Z [verbose] multus-daemon started\\\\n2025-12-03T00:07:17Z [verbose] Readiness Indicator file check\\\\n2025-12-03T00:08:02Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T00:07:16Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:08:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gdrw4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:07:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-675qk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:08:14Z is after 2025-08-24T17:21:41Z" Dec 03 00:08:15 crc kubenswrapper[4953]: I1203 00:08:15.010802 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f94f8c51-c665-4ca3-bf17-7a04ea5c7ade\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72c5cd7dc1c3a8adfedab6ba43a59d80b3efe65f2bb56be95e76cc86a409dcf3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2810ffc1b141fa7da811406053db9d72c708edf5f2bb1dd75e4d52ab35fd24c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c67b4277b20a9f1287f50ef82d0785c9d02543a8e7ed26e0413eae332b866db7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://258d78e1ad9d3dcda6976909ae3fe71aecead5b995fc66b3fd6e2722d80b5ab8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:06:51Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:08:15Z is after 2025-08-24T17:21:41Z" Dec 03 00:08:15 crc kubenswrapper[4953]: I1203 00:08:15.025548 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5652d0f3-3fd7-4af6-8941-9119e1177641\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2030aef5b7170991268fa5dc7e7ed8ce1fbe3a2765165a10c25251426575330f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ab6052f1abf17225155ce1bd86afa742e6e78f507b7deedd7161cd8fba9120b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://650b6c79cda2e1768ba1d8763db799e210ce2748893512068d25e60f821937ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d1d60162f7f6885d4f51b4ab44163cfae6d2d04355b1dd710e26b022450005df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d1d60162f7f6885d4f51b4ab44163cfae6d2d04355b1dd710e26b022450005df\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:06:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:06:51Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:08:15Z is after 2025-08-24T17:21:41Z" Dec 03 00:08:15 crc kubenswrapper[4953]: I1203 00:08:15.033922 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:15 crc kubenswrapper[4953]: I1203 00:08:15.033988 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:15 crc kubenswrapper[4953]: I1203 00:08:15.034010 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:15 crc kubenswrapper[4953]: I1203 00:08:15.034031 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:15 crc kubenswrapper[4953]: I1203 00:08:15.034046 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:15Z","lastTransitionTime":"2025-12-03T00:08:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:15 crc kubenswrapper[4953]: I1203 00:08:15.041306 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09657a9f448a4c92a759c73a0a892076d7eb46393e6fd565e5f73f34c71aac23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5af5eac4a58870f63497f99b503cfa5fb922278095e0332ec50376925ecf2109\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:08:15Z is after 2025-08-24T17:21:41Z" Dec 03 00:08:15 crc kubenswrapper[4953]: I1203 00:08:15.078316 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-wj6rz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"96aa67a0-15dd-420d-a1d5-ed239d7df1fb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01815cfd3f58c3f3f5f2a894c1790e0b69c5060c48dc0b1deff050b6ad7af916\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kf62m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b12bfb5bdf9fd4f2c50a335fc82df6ee0440eaf8741897da03aa2dae368cc6b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b12bfb5bdf9fd4f2c50a335fc82df6ee0440eaf8741897da03aa2dae368cc6b4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:07:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kf62m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83aaa5fa28110bc0ef05aff75cd1442bc9bbdd33b822ae2ef22d4e4ab83238bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://83aaa5fa28110bc0ef05aff75cd1442bc9bbdd33b822ae2ef22d4e4ab83238bc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:07:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:07:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kf62m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ab6213a33a19fedebce6d74a0ebbb3b3a2de5687504e21cc24093cc42ade635\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ab6213a33a19fedebce6d74a0ebbb3b3a2de5687504e21cc24093cc42ade635\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:07:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:07:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kf62m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://de480488843afa6b8dede1e1b42b5181dca1993e2f39dc1b461a18a57e3668e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://de480488843afa6b8dede1e1b42b5181dca1993e2f39dc1b461a18a57e3668e1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:07:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kf62m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://97be4ba692bc61078d4b0ba39fb0c12772548c6c9251ff3356ea702f151223cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://97be4ba692bc61078d4b0ba39fb0c12772548c6c9251ff3356ea702f151223cd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:07:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kf62m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1a1f009fcbd8af9217db1dc97f662719f8c1db232df0e2804e44c881abb1df2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b1a1f009fcbd8af9217db1dc97f662719f8c1db232df0e2804e44c881abb1df2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:07:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:07:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kf62m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:07:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-wj6rz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:08:15Z is after 2025-08-24T17:21:41Z" Dec 03 00:08:15 crc kubenswrapper[4953]: I1203 00:08:15.099995 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-kl5s9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d85e5c97-16ad-4bd7-9e73-b7f12678be97\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c21df91a9e4e3165686882b2eb559498224063dc633c3a63f958598108eca29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j24lb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:07:16Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-kl5s9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:08:15Z is after 2025-08-24T17:21:41Z" Dec 03 00:08:15 crc kubenswrapper[4953]: I1203 00:08:15.115102 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-chmnz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a12fe6d-d25a-4b08-90da-cc9e510980a7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://711ab839cb9c0a90554dadbd45d810fc78aac09739340482fef8de8743b77283\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c982g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f802d5628a8cc32618044b9383211d5a492c56060d427f953b7793a1064029e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c982g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:07:29Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-chmnz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:08:15Z is after 2025-08-24T17:21:41Z" Dec 03 00:08:15 crc kubenswrapper[4953]: I1203 00:08:15.126246 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-xshk4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bd03a758-0ca9-47d8-86f7-18b2d610df43\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47gfh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47gfh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:07:31Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-xshk4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:08:15Z is after 2025-08-24T17:21:41Z" Dec 03 00:08:15 crc kubenswrapper[4953]: I1203 00:08:15.136463 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:15 crc kubenswrapper[4953]: I1203 00:08:15.136667 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:15 crc kubenswrapper[4953]: I1203 00:08:15.136761 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:15 crc kubenswrapper[4953]: I1203 00:08:15.136869 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:15 crc kubenswrapper[4953]: I1203 00:08:15.136991 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:15Z","lastTransitionTime":"2025-12-03T00:08:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:15 crc kubenswrapper[4953]: I1203 00:08:15.139399 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cf821840-5353-4620-8f73-99b56b371ab0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cae5e2994fcc24dc9b2ea73e79572615b3311edc42642b7f278fb84123989b23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://29e4b9d1b70a93fb133e0759ad148ae133766a66efeae3ff06a21dfae0e805ff\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dec43536772aa7c3f5a662fa7f3bd73b1cae2791ad454f85792287828d0ffeed\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://034e0bac6892d361700b8f33d6427b8a55e2bd3f89b14ea77a374c6be94d25e2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://761973393a93c59c23e69f41b8ac0d16abc35abd61ecbc6d51a42d94cd664233\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T00:07:09Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 00:07:03.713374 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 00:07:03.715052 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1528563048/tls.crt::/tmp/serving-cert-1528563048/tls.key\\\\\\\"\\\\nI1203 00:07:09.300052 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 00:07:09.304192 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 00:07:09.304255 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 00:07:09.304309 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 00:07:09.304323 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 00:07:09.313563 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 00:07:09.313597 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 00:07:09.313603 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 00:07:09.313609 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 00:07:09.313613 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 00:07:09.313618 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 00:07:09.313622 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 00:07:09.313709 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 00:07:09.327347 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T00:06:53Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c143d4232fb8bf7ef5999a664e431a00ab0db030f4d31a28d0bb21e00e38581\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:53Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72af6c27296eae2e275d455795c68f0d272869d63642b8353b17d466d0c901cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://72af6c27296eae2e275d455795c68f0d272869d63642b8353b17d466d0c901cf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:06:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:06:51Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:08:15Z is after 2025-08-24T17:21:41Z" Dec 03 00:08:15 crc kubenswrapper[4953]: I1203 00:08:15.152144 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:10Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:08:15Z is after 2025-08-24T17:21:41Z" Dec 03 00:08:15 crc kubenswrapper[4953]: I1203 00:08:15.162462 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce8b414e1fc22eb30decb6a6411b672e7f443055cb89ebb029fbf6ffc1d5dca7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:08:15Z is after 2025-08-24T17:21:41Z" Dec 03 00:08:15 crc kubenswrapper[4953]: I1203 00:08:15.173882 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-6dfbn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53ffb1e9-9a2e-4869-80d7-776f4aa7ef8b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ab4205c0d38ad6a0c32077eadb1fd5fd3a279d9c8b8a2ffbe693ec2c4018370c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ct26c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7df84405eda520ad5a22d35750edf07ef3c3977a756ad5739c54f91cba9da7b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ct26c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:07:16Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-6dfbn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:08:15Z is after 2025-08-24T17:21:41Z" Dec 03 00:08:15 crc kubenswrapper[4953]: I1203 00:08:15.192278 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-txffx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8816ebee-c735-4b8d-b1ec-65349707b347\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:16Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:16Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4b85224419a4075acb340e4b11095f2adb5c316e285afa4e19295c2181c0d24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0f6f645a58993f052a6f365bb808f62d99f7fea74886c5cc022b2d86112321a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a61b951a9cbab2c48c0f2df8351f97d7ff71bf3674f641fe5eac30155232b224\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd5f428de229749485267e81d34e31b708728b52474032e88460e54747f21e11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://00a2ea4a9bb870fce30e7fa845f787f184e7a19e42c95511655ea8c5f84a26d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc0a5d4cc9d614d8492626a1b114bdc52f510eb5965d842c02c083f08d4849a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a4b625eda9b9daff8aa7fe3c5b94e417bcefe08fc6a65b4594f821cfdca7bfd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0a4b625eda9b9daff8aa7fe3c5b94e417bcefe08fc6a65b4594f821cfdca7bfd\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T00:08:13Z\\\",\\\"message\\\":\\\"il\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1203 00:08:13.521854 6989 transact.go:42] Configuring OVN: [{Op:update Table:Logical_Router_Static_Route Row:map[ip_prefix:10.217.0.0/22 nexthop:100.64.0.2 policy:{GoSet:[src-ip]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {8944024f-deb7-4076-afb3-4b50a2ff4b4b}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:mutate Table:Logical_Router Row:map[] Rows:[] Columns:[] Mutations:[{Column:static_routes Mutator:insert Value:{GoSet:[{GoUUID:8944024f-deb7-4076-afb3-4b50a2ff4b4b}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {f6d604c1-9711-4e25-be6c-79ec28bbad1b}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1203 00:08:13.522171 6989 obj_retry.go:551] Creating *factory.egressNode crc took: 1.771964ms\\\\nI1203 00:08:13.522207 6989 factory.go:1336] Added *v1.Node event handler 7\\\\nI1203 00:08:13.522242 6989 factory.go:1336] Added *v1.EgressIP event handler 8\\\\nI1203 00:08:13.522575 6989 factory.go:1336] Added *v1.EgressFirewall event handler 9\\\\nI1203 00:08:13.522717 6989 controller.go:132] Adding controller ef_node_controller event handlers\\\\nI1203 00:08:13.522757 6989 ovnkube.go:599] Stopped ovnkube\\\\nI1203 00:08:13.522794 6989 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1203 00:08:13.522871 6989 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T00:08:12Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-txffx_openshift-ovn-kubernetes(8816ebee-c735-4b8d-b1ec-65349707b347)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7deadc8e135ff1bc7c5a2f093e489bb5b4b6f166bc874c6acc2a841c5397f11e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af4c7764fca87144d01f3738ea010308744ccf50c54b7c471ccc213112de00d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af4c7764fca87144d01f3738ea010308744ccf50c54b7c471ccc213112de00d5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:07:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:07:16Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-txffx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:08:15Z is after 2025-08-24T17:21:41Z" Dec 03 00:08:15 crc kubenswrapper[4953]: I1203 00:08:15.202088 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5cmlc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7e11a1d3-ecee-4bf0-87bf-085fab97b73d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d42390b9c53efc63add8ddcebf03c976f5e5448eaa2c34e076b2dc90dbac60fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6x2sc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:07:18Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5cmlc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:08:15Z is after 2025-08-24T17:21:41Z" Dec 03 00:08:15 crc kubenswrapper[4953]: I1203 00:08:15.239844 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:15 crc kubenswrapper[4953]: I1203 00:08:15.239870 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:15 crc kubenswrapper[4953]: I1203 00:08:15.239880 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:15 crc kubenswrapper[4953]: I1203 00:08:15.239896 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:15 crc kubenswrapper[4953]: I1203 00:08:15.239908 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:15Z","lastTransitionTime":"2025-12-03T00:08:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:15 crc kubenswrapper[4953]: I1203 00:08:15.342507 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:15 crc kubenswrapper[4953]: I1203 00:08:15.342552 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:15 crc kubenswrapper[4953]: I1203 00:08:15.342564 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:15 crc kubenswrapper[4953]: I1203 00:08:15.342580 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:15 crc kubenswrapper[4953]: I1203 00:08:15.342593 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:15Z","lastTransitionTime":"2025-12-03T00:08:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:15 crc kubenswrapper[4953]: I1203 00:08:15.347660 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xshk4" Dec 03 00:08:15 crc kubenswrapper[4953]: E1203 00:08:15.347954 4953 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xshk4" podUID="bd03a758-0ca9-47d8-86f7-18b2d610df43" Dec 03 00:08:15 crc kubenswrapper[4953]: I1203 00:08:15.449719 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:15 crc kubenswrapper[4953]: I1203 00:08:15.450027 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:15 crc kubenswrapper[4953]: I1203 00:08:15.450042 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:15 crc kubenswrapper[4953]: I1203 00:08:15.450060 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:15 crc kubenswrapper[4953]: I1203 00:08:15.450455 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:15Z","lastTransitionTime":"2025-12-03T00:08:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:15 crc kubenswrapper[4953]: I1203 00:08:15.553038 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:15 crc kubenswrapper[4953]: I1203 00:08:15.553112 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:15 crc kubenswrapper[4953]: I1203 00:08:15.553137 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:15 crc kubenswrapper[4953]: I1203 00:08:15.553168 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:15 crc kubenswrapper[4953]: I1203 00:08:15.553192 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:15Z","lastTransitionTime":"2025-12-03T00:08:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:15 crc kubenswrapper[4953]: I1203 00:08:15.656148 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:15 crc kubenswrapper[4953]: I1203 00:08:15.656200 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:15 crc kubenswrapper[4953]: I1203 00:08:15.656216 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:15 crc kubenswrapper[4953]: I1203 00:08:15.656238 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:15 crc kubenswrapper[4953]: I1203 00:08:15.656255 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:15Z","lastTransitionTime":"2025-12-03T00:08:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:15 crc kubenswrapper[4953]: I1203 00:08:15.758839 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:15 crc kubenswrapper[4953]: I1203 00:08:15.758940 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:15 crc kubenswrapper[4953]: I1203 00:08:15.758960 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:15 crc kubenswrapper[4953]: I1203 00:08:15.759016 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:15 crc kubenswrapper[4953]: I1203 00:08:15.759036 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:15Z","lastTransitionTime":"2025-12-03T00:08:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:15 crc kubenswrapper[4953]: I1203 00:08:15.861935 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:15 crc kubenswrapper[4953]: I1203 00:08:15.862024 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:15 crc kubenswrapper[4953]: I1203 00:08:15.862035 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:15 crc kubenswrapper[4953]: I1203 00:08:15.862057 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:15 crc kubenswrapper[4953]: I1203 00:08:15.862070 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:15Z","lastTransitionTime":"2025-12-03T00:08:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:15 crc kubenswrapper[4953]: I1203 00:08:15.965018 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:15 crc kubenswrapper[4953]: I1203 00:08:15.965082 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:15 crc kubenswrapper[4953]: I1203 00:08:15.965100 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:15 crc kubenswrapper[4953]: I1203 00:08:15.965130 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:15 crc kubenswrapper[4953]: I1203 00:08:15.965152 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:15Z","lastTransitionTime":"2025-12-03T00:08:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:16 crc kubenswrapper[4953]: I1203 00:08:16.068073 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:16 crc kubenswrapper[4953]: I1203 00:08:16.068127 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:16 crc kubenswrapper[4953]: I1203 00:08:16.068137 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:16 crc kubenswrapper[4953]: I1203 00:08:16.068156 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:16 crc kubenswrapper[4953]: I1203 00:08:16.068167 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:16Z","lastTransitionTime":"2025-12-03T00:08:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:16 crc kubenswrapper[4953]: I1203 00:08:16.170956 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:16 crc kubenswrapper[4953]: I1203 00:08:16.171130 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:16 crc kubenswrapper[4953]: I1203 00:08:16.171161 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:16 crc kubenswrapper[4953]: I1203 00:08:16.171193 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:16 crc kubenswrapper[4953]: I1203 00:08:16.171215 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:16Z","lastTransitionTime":"2025-12-03T00:08:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:16 crc kubenswrapper[4953]: I1203 00:08:16.275060 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:16 crc kubenswrapper[4953]: I1203 00:08:16.275591 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:16 crc kubenswrapper[4953]: I1203 00:08:16.275766 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:16 crc kubenswrapper[4953]: I1203 00:08:16.275951 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:16 crc kubenswrapper[4953]: I1203 00:08:16.276145 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:16Z","lastTransitionTime":"2025-12-03T00:08:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:16 crc kubenswrapper[4953]: I1203 00:08:16.347616 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 00:08:16 crc kubenswrapper[4953]: I1203 00:08:16.347633 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 00:08:16 crc kubenswrapper[4953]: E1203 00:08:16.347828 4953 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 00:08:16 crc kubenswrapper[4953]: I1203 00:08:16.347744 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 00:08:16 crc kubenswrapper[4953]: E1203 00:08:16.347893 4953 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 00:08:16 crc kubenswrapper[4953]: E1203 00:08:16.348156 4953 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 00:08:16 crc kubenswrapper[4953]: I1203 00:08:16.378582 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:16 crc kubenswrapper[4953]: I1203 00:08:16.378934 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:16 crc kubenswrapper[4953]: I1203 00:08:16.379110 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:16 crc kubenswrapper[4953]: I1203 00:08:16.379353 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:16 crc kubenswrapper[4953]: I1203 00:08:16.379541 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:16Z","lastTransitionTime":"2025-12-03T00:08:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:16 crc kubenswrapper[4953]: I1203 00:08:16.483864 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:16 crc kubenswrapper[4953]: I1203 00:08:16.484471 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:16 crc kubenswrapper[4953]: I1203 00:08:16.484609 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:16 crc kubenswrapper[4953]: I1203 00:08:16.484699 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:16 crc kubenswrapper[4953]: I1203 00:08:16.484798 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:16Z","lastTransitionTime":"2025-12-03T00:08:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:16 crc kubenswrapper[4953]: I1203 00:08:16.587882 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:16 crc kubenswrapper[4953]: I1203 00:08:16.588330 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:16 crc kubenswrapper[4953]: I1203 00:08:16.588480 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:16 crc kubenswrapper[4953]: I1203 00:08:16.588900 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:16 crc kubenswrapper[4953]: I1203 00:08:16.589202 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:16Z","lastTransitionTime":"2025-12-03T00:08:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:16 crc kubenswrapper[4953]: I1203 00:08:16.692629 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:16 crc kubenswrapper[4953]: I1203 00:08:16.692697 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:16 crc kubenswrapper[4953]: I1203 00:08:16.692714 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:16 crc kubenswrapper[4953]: I1203 00:08:16.692740 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:16 crc kubenswrapper[4953]: I1203 00:08:16.692758 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:16Z","lastTransitionTime":"2025-12-03T00:08:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:16 crc kubenswrapper[4953]: I1203 00:08:16.796418 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:16 crc kubenswrapper[4953]: I1203 00:08:16.796491 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:16 crc kubenswrapper[4953]: I1203 00:08:16.796522 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:16 crc kubenswrapper[4953]: I1203 00:08:16.796557 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:16 crc kubenswrapper[4953]: I1203 00:08:16.796582 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:16Z","lastTransitionTime":"2025-12-03T00:08:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:16 crc kubenswrapper[4953]: I1203 00:08:16.899333 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:16 crc kubenswrapper[4953]: I1203 00:08:16.899454 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:16 crc kubenswrapper[4953]: I1203 00:08:16.899485 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:16 crc kubenswrapper[4953]: I1203 00:08:16.899521 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:16 crc kubenswrapper[4953]: I1203 00:08:16.899545 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:16Z","lastTransitionTime":"2025-12-03T00:08:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:17 crc kubenswrapper[4953]: I1203 00:08:17.003949 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:17 crc kubenswrapper[4953]: I1203 00:08:17.004743 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:17 crc kubenswrapper[4953]: I1203 00:08:17.004933 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:17 crc kubenswrapper[4953]: I1203 00:08:17.005161 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:17 crc kubenswrapper[4953]: I1203 00:08:17.005316 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:17Z","lastTransitionTime":"2025-12-03T00:08:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:17 crc kubenswrapper[4953]: I1203 00:08:17.109113 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:17 crc kubenswrapper[4953]: I1203 00:08:17.109168 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:17 crc kubenswrapper[4953]: I1203 00:08:17.109184 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:17 crc kubenswrapper[4953]: I1203 00:08:17.109206 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:17 crc kubenswrapper[4953]: I1203 00:08:17.109223 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:17Z","lastTransitionTime":"2025-12-03T00:08:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:17 crc kubenswrapper[4953]: I1203 00:08:17.212595 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:17 crc kubenswrapper[4953]: I1203 00:08:17.212672 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:17 crc kubenswrapper[4953]: I1203 00:08:17.212694 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:17 crc kubenswrapper[4953]: I1203 00:08:17.212727 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:17 crc kubenswrapper[4953]: I1203 00:08:17.212749 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:17Z","lastTransitionTime":"2025-12-03T00:08:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:17 crc kubenswrapper[4953]: I1203 00:08:17.315676 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:17 crc kubenswrapper[4953]: I1203 00:08:17.315741 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:17 crc kubenswrapper[4953]: I1203 00:08:17.315761 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:17 crc kubenswrapper[4953]: I1203 00:08:17.315787 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:17 crc kubenswrapper[4953]: I1203 00:08:17.315805 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:17Z","lastTransitionTime":"2025-12-03T00:08:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:17 crc kubenswrapper[4953]: I1203 00:08:17.348051 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xshk4" Dec 03 00:08:17 crc kubenswrapper[4953]: E1203 00:08:17.348579 4953 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xshk4" podUID="bd03a758-0ca9-47d8-86f7-18b2d610df43" Dec 03 00:08:17 crc kubenswrapper[4953]: I1203 00:08:17.418216 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:17 crc kubenswrapper[4953]: I1203 00:08:17.418320 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:17 crc kubenswrapper[4953]: I1203 00:08:17.418353 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:17 crc kubenswrapper[4953]: I1203 00:08:17.418387 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:17 crc kubenswrapper[4953]: I1203 00:08:17.418411 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:17Z","lastTransitionTime":"2025-12-03T00:08:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:17 crc kubenswrapper[4953]: I1203 00:08:17.521927 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:17 crc kubenswrapper[4953]: I1203 00:08:17.522057 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:17 crc kubenswrapper[4953]: I1203 00:08:17.522135 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:17 crc kubenswrapper[4953]: I1203 00:08:17.522183 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:17 crc kubenswrapper[4953]: I1203 00:08:17.522250 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:17Z","lastTransitionTime":"2025-12-03T00:08:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:17 crc kubenswrapper[4953]: I1203 00:08:17.625078 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:17 crc kubenswrapper[4953]: I1203 00:08:17.625138 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:17 crc kubenswrapper[4953]: I1203 00:08:17.625155 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:17 crc kubenswrapper[4953]: I1203 00:08:17.625187 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:17 crc kubenswrapper[4953]: I1203 00:08:17.625223 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:17Z","lastTransitionTime":"2025-12-03T00:08:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:17 crc kubenswrapper[4953]: I1203 00:08:17.728272 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:17 crc kubenswrapper[4953]: I1203 00:08:17.728509 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:17 crc kubenswrapper[4953]: I1203 00:08:17.728574 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:17 crc kubenswrapper[4953]: I1203 00:08:17.728638 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:17 crc kubenswrapper[4953]: I1203 00:08:17.728726 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:17Z","lastTransitionTime":"2025-12-03T00:08:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:17 crc kubenswrapper[4953]: I1203 00:08:17.831717 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:17 crc kubenswrapper[4953]: I1203 00:08:17.831776 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:17 crc kubenswrapper[4953]: I1203 00:08:17.831792 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:17 crc kubenswrapper[4953]: I1203 00:08:17.831854 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:17 crc kubenswrapper[4953]: I1203 00:08:17.831871 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:17Z","lastTransitionTime":"2025-12-03T00:08:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:17 crc kubenswrapper[4953]: I1203 00:08:17.934360 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:17 crc kubenswrapper[4953]: I1203 00:08:17.934779 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:17 crc kubenswrapper[4953]: I1203 00:08:17.934954 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:17 crc kubenswrapper[4953]: I1203 00:08:17.935182 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:17 crc kubenswrapper[4953]: I1203 00:08:17.935362 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:17Z","lastTransitionTime":"2025-12-03T00:08:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:18 crc kubenswrapper[4953]: I1203 00:08:18.038382 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:18 crc kubenswrapper[4953]: I1203 00:08:18.038446 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:18 crc kubenswrapper[4953]: I1203 00:08:18.038467 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:18 crc kubenswrapper[4953]: I1203 00:08:18.038498 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:18 crc kubenswrapper[4953]: I1203 00:08:18.038522 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:18Z","lastTransitionTime":"2025-12-03T00:08:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:18 crc kubenswrapper[4953]: I1203 00:08:18.141353 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:18 crc kubenswrapper[4953]: I1203 00:08:18.141419 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:18 crc kubenswrapper[4953]: I1203 00:08:18.141445 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:18 crc kubenswrapper[4953]: I1203 00:08:18.141475 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:18 crc kubenswrapper[4953]: I1203 00:08:18.141499 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:18Z","lastTransitionTime":"2025-12-03T00:08:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:18 crc kubenswrapper[4953]: I1203 00:08:18.248394 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:18 crc kubenswrapper[4953]: I1203 00:08:18.248485 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:18 crc kubenswrapper[4953]: I1203 00:08:18.248510 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:18 crc kubenswrapper[4953]: I1203 00:08:18.248544 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:18 crc kubenswrapper[4953]: I1203 00:08:18.248579 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:18Z","lastTransitionTime":"2025-12-03T00:08:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:18 crc kubenswrapper[4953]: I1203 00:08:18.347266 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 00:08:18 crc kubenswrapper[4953]: E1203 00:08:18.347763 4953 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 00:08:18 crc kubenswrapper[4953]: I1203 00:08:18.347393 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 00:08:18 crc kubenswrapper[4953]: E1203 00:08:18.348301 4953 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 00:08:18 crc kubenswrapper[4953]: I1203 00:08:18.347267 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 00:08:18 crc kubenswrapper[4953]: E1203 00:08:18.348673 4953 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 00:08:18 crc kubenswrapper[4953]: I1203 00:08:18.351410 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:18 crc kubenswrapper[4953]: I1203 00:08:18.351469 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:18 crc kubenswrapper[4953]: I1203 00:08:18.351486 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:18 crc kubenswrapper[4953]: I1203 00:08:18.351512 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:18 crc kubenswrapper[4953]: I1203 00:08:18.351531 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:18Z","lastTransitionTime":"2025-12-03T00:08:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:18 crc kubenswrapper[4953]: I1203 00:08:18.454526 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:18 crc kubenswrapper[4953]: I1203 00:08:18.454603 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:18 crc kubenswrapper[4953]: I1203 00:08:18.454625 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:18 crc kubenswrapper[4953]: I1203 00:08:18.454671 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:18 crc kubenswrapper[4953]: I1203 00:08:18.454688 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:18Z","lastTransitionTime":"2025-12-03T00:08:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:18 crc kubenswrapper[4953]: I1203 00:08:18.557435 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:18 crc kubenswrapper[4953]: I1203 00:08:18.557524 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:18 crc kubenswrapper[4953]: I1203 00:08:18.557540 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:18 crc kubenswrapper[4953]: I1203 00:08:18.557563 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:18 crc kubenswrapper[4953]: I1203 00:08:18.557581 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:18Z","lastTransitionTime":"2025-12-03T00:08:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:18 crc kubenswrapper[4953]: I1203 00:08:18.661222 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:18 crc kubenswrapper[4953]: I1203 00:08:18.661577 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:18 crc kubenswrapper[4953]: I1203 00:08:18.661797 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:18 crc kubenswrapper[4953]: I1203 00:08:18.662043 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:18 crc kubenswrapper[4953]: I1203 00:08:18.662273 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:18Z","lastTransitionTime":"2025-12-03T00:08:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:18 crc kubenswrapper[4953]: I1203 00:08:18.764717 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:18 crc kubenswrapper[4953]: I1203 00:08:18.764778 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:18 crc kubenswrapper[4953]: I1203 00:08:18.764797 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:18 crc kubenswrapper[4953]: I1203 00:08:18.764821 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:18 crc kubenswrapper[4953]: I1203 00:08:18.764838 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:18Z","lastTransitionTime":"2025-12-03T00:08:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:18 crc kubenswrapper[4953]: I1203 00:08:18.868199 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:18 crc kubenswrapper[4953]: I1203 00:08:18.868682 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:18 crc kubenswrapper[4953]: I1203 00:08:18.868859 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:18 crc kubenswrapper[4953]: I1203 00:08:18.869054 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:18 crc kubenswrapper[4953]: I1203 00:08:18.869252 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:18Z","lastTransitionTime":"2025-12-03T00:08:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:18 crc kubenswrapper[4953]: I1203 00:08:18.973457 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:18 crc kubenswrapper[4953]: I1203 00:08:18.973914 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:18 crc kubenswrapper[4953]: I1203 00:08:18.974233 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:18 crc kubenswrapper[4953]: I1203 00:08:18.974424 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:18 crc kubenswrapper[4953]: I1203 00:08:18.974577 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:18Z","lastTransitionTime":"2025-12-03T00:08:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:19 crc kubenswrapper[4953]: I1203 00:08:19.077623 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:19 crc kubenswrapper[4953]: I1203 00:08:19.077935 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:19 crc kubenswrapper[4953]: I1203 00:08:19.078311 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:19 crc kubenswrapper[4953]: I1203 00:08:19.078483 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:19 crc kubenswrapper[4953]: I1203 00:08:19.078618 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:19Z","lastTransitionTime":"2025-12-03T00:08:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:19 crc kubenswrapper[4953]: I1203 00:08:19.182309 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:19 crc kubenswrapper[4953]: I1203 00:08:19.182700 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:19 crc kubenswrapper[4953]: I1203 00:08:19.182849 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:19 crc kubenswrapper[4953]: I1203 00:08:19.183022 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:19 crc kubenswrapper[4953]: I1203 00:08:19.183201 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:19Z","lastTransitionTime":"2025-12-03T00:08:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:19 crc kubenswrapper[4953]: I1203 00:08:19.286386 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:19 crc kubenswrapper[4953]: I1203 00:08:19.286446 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:19 crc kubenswrapper[4953]: I1203 00:08:19.286463 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:19 crc kubenswrapper[4953]: I1203 00:08:19.286485 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:19 crc kubenswrapper[4953]: I1203 00:08:19.286500 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:19Z","lastTransitionTime":"2025-12-03T00:08:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:19 crc kubenswrapper[4953]: I1203 00:08:19.347585 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xshk4" Dec 03 00:08:19 crc kubenswrapper[4953]: E1203 00:08:19.347769 4953 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xshk4" podUID="bd03a758-0ca9-47d8-86f7-18b2d610df43" Dec 03 00:08:19 crc kubenswrapper[4953]: I1203 00:08:19.390233 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:19 crc kubenswrapper[4953]: I1203 00:08:19.391268 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:19 crc kubenswrapper[4953]: I1203 00:08:19.391487 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:19 crc kubenswrapper[4953]: I1203 00:08:19.391759 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:19 crc kubenswrapper[4953]: I1203 00:08:19.392006 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:19Z","lastTransitionTime":"2025-12-03T00:08:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:19 crc kubenswrapper[4953]: I1203 00:08:19.495603 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:19 crc kubenswrapper[4953]: I1203 00:08:19.495672 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:19 crc kubenswrapper[4953]: I1203 00:08:19.495692 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:19 crc kubenswrapper[4953]: I1203 00:08:19.495717 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:19 crc kubenswrapper[4953]: I1203 00:08:19.495737 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:19Z","lastTransitionTime":"2025-12-03T00:08:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:19 crc kubenswrapper[4953]: I1203 00:08:19.599116 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:19 crc kubenswrapper[4953]: I1203 00:08:19.599222 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:19 crc kubenswrapper[4953]: I1203 00:08:19.599240 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:19 crc kubenswrapper[4953]: I1203 00:08:19.599272 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:19 crc kubenswrapper[4953]: I1203 00:08:19.599295 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:19Z","lastTransitionTime":"2025-12-03T00:08:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:19 crc kubenswrapper[4953]: I1203 00:08:19.701816 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:19 crc kubenswrapper[4953]: I1203 00:08:19.701916 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:19 crc kubenswrapper[4953]: I1203 00:08:19.701929 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:19 crc kubenswrapper[4953]: I1203 00:08:19.701949 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:19 crc kubenswrapper[4953]: I1203 00:08:19.701962 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:19Z","lastTransitionTime":"2025-12-03T00:08:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:19 crc kubenswrapper[4953]: I1203 00:08:19.804862 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:19 crc kubenswrapper[4953]: I1203 00:08:19.804915 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:19 crc kubenswrapper[4953]: I1203 00:08:19.804925 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:19 crc kubenswrapper[4953]: I1203 00:08:19.804940 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:19 crc kubenswrapper[4953]: I1203 00:08:19.804950 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:19Z","lastTransitionTime":"2025-12-03T00:08:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:19 crc kubenswrapper[4953]: I1203 00:08:19.907591 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:19 crc kubenswrapper[4953]: I1203 00:08:19.907660 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:19 crc kubenswrapper[4953]: I1203 00:08:19.907679 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:19 crc kubenswrapper[4953]: I1203 00:08:19.907701 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:19 crc kubenswrapper[4953]: I1203 00:08:19.907720 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:19Z","lastTransitionTime":"2025-12-03T00:08:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:20 crc kubenswrapper[4953]: I1203 00:08:20.010345 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:20 crc kubenswrapper[4953]: I1203 00:08:20.010405 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:20 crc kubenswrapper[4953]: I1203 00:08:20.010423 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:20 crc kubenswrapper[4953]: I1203 00:08:20.010446 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:20 crc kubenswrapper[4953]: I1203 00:08:20.010465 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:20Z","lastTransitionTime":"2025-12-03T00:08:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:20 crc kubenswrapper[4953]: I1203 00:08:20.113028 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:20 crc kubenswrapper[4953]: I1203 00:08:20.113112 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:20 crc kubenswrapper[4953]: I1203 00:08:20.113130 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:20 crc kubenswrapper[4953]: I1203 00:08:20.113157 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:20 crc kubenswrapper[4953]: I1203 00:08:20.113175 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:20Z","lastTransitionTime":"2025-12-03T00:08:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:20 crc kubenswrapper[4953]: I1203 00:08:20.215438 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:20 crc kubenswrapper[4953]: I1203 00:08:20.215480 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:20 crc kubenswrapper[4953]: I1203 00:08:20.215491 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:20 crc kubenswrapper[4953]: I1203 00:08:20.215508 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:20 crc kubenswrapper[4953]: I1203 00:08:20.215522 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:20Z","lastTransitionTime":"2025-12-03T00:08:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:20 crc kubenswrapper[4953]: I1203 00:08:20.317856 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:20 crc kubenswrapper[4953]: I1203 00:08:20.317887 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:20 crc kubenswrapper[4953]: I1203 00:08:20.317897 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:20 crc kubenswrapper[4953]: I1203 00:08:20.317911 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:20 crc kubenswrapper[4953]: I1203 00:08:20.317921 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:20Z","lastTransitionTime":"2025-12-03T00:08:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:20 crc kubenswrapper[4953]: I1203 00:08:20.347967 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 00:08:20 crc kubenswrapper[4953]: I1203 00:08:20.348082 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 00:08:20 crc kubenswrapper[4953]: E1203 00:08:20.348236 4953 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 00:08:20 crc kubenswrapper[4953]: I1203 00:08:20.348277 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 00:08:20 crc kubenswrapper[4953]: E1203 00:08:20.348457 4953 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 00:08:20 crc kubenswrapper[4953]: E1203 00:08:20.348578 4953 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 00:08:20 crc kubenswrapper[4953]: I1203 00:08:20.420728 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:20 crc kubenswrapper[4953]: I1203 00:08:20.420779 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:20 crc kubenswrapper[4953]: I1203 00:08:20.420797 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:20 crc kubenswrapper[4953]: I1203 00:08:20.420821 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:20 crc kubenswrapper[4953]: I1203 00:08:20.420838 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:20Z","lastTransitionTime":"2025-12-03T00:08:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:20 crc kubenswrapper[4953]: I1203 00:08:20.523718 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:20 crc kubenswrapper[4953]: I1203 00:08:20.523794 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:20 crc kubenswrapper[4953]: I1203 00:08:20.523911 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:20 crc kubenswrapper[4953]: I1203 00:08:20.523939 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:20 crc kubenswrapper[4953]: I1203 00:08:20.523956 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:20Z","lastTransitionTime":"2025-12-03T00:08:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:20 crc kubenswrapper[4953]: I1203 00:08:20.627037 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:20 crc kubenswrapper[4953]: I1203 00:08:20.627091 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:20 crc kubenswrapper[4953]: I1203 00:08:20.627107 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:20 crc kubenswrapper[4953]: I1203 00:08:20.627130 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:20 crc kubenswrapper[4953]: I1203 00:08:20.627149 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:20Z","lastTransitionTime":"2025-12-03T00:08:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:20 crc kubenswrapper[4953]: I1203 00:08:20.730183 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:20 crc kubenswrapper[4953]: I1203 00:08:20.730243 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:20 crc kubenswrapper[4953]: I1203 00:08:20.730265 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:20 crc kubenswrapper[4953]: I1203 00:08:20.730293 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:20 crc kubenswrapper[4953]: I1203 00:08:20.730314 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:20Z","lastTransitionTime":"2025-12-03T00:08:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:20 crc kubenswrapper[4953]: I1203 00:08:20.833761 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:20 crc kubenswrapper[4953]: I1203 00:08:20.833825 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:20 crc kubenswrapper[4953]: I1203 00:08:20.833848 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:20 crc kubenswrapper[4953]: I1203 00:08:20.833877 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:20 crc kubenswrapper[4953]: I1203 00:08:20.833900 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:20Z","lastTransitionTime":"2025-12-03T00:08:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:20 crc kubenswrapper[4953]: I1203 00:08:20.936571 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:20 crc kubenswrapper[4953]: I1203 00:08:20.936636 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:20 crc kubenswrapper[4953]: I1203 00:08:20.936657 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:20 crc kubenswrapper[4953]: I1203 00:08:20.936685 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:20 crc kubenswrapper[4953]: I1203 00:08:20.936705 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:20Z","lastTransitionTime":"2025-12-03T00:08:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:21 crc kubenswrapper[4953]: I1203 00:08:21.040277 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:21 crc kubenswrapper[4953]: I1203 00:08:21.040347 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:21 crc kubenswrapper[4953]: I1203 00:08:21.040362 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:21 crc kubenswrapper[4953]: I1203 00:08:21.040406 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:21 crc kubenswrapper[4953]: I1203 00:08:21.040423 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:21Z","lastTransitionTime":"2025-12-03T00:08:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:21 crc kubenswrapper[4953]: I1203 00:08:21.143637 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:21 crc kubenswrapper[4953]: I1203 00:08:21.143711 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:21 crc kubenswrapper[4953]: I1203 00:08:21.143737 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:21 crc kubenswrapper[4953]: I1203 00:08:21.143763 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:21 crc kubenswrapper[4953]: I1203 00:08:21.143785 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:21Z","lastTransitionTime":"2025-12-03T00:08:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:21 crc kubenswrapper[4953]: I1203 00:08:21.246662 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:21 crc kubenswrapper[4953]: I1203 00:08:21.246696 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:21 crc kubenswrapper[4953]: I1203 00:08:21.246705 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:21 crc kubenswrapper[4953]: I1203 00:08:21.246718 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:21 crc kubenswrapper[4953]: I1203 00:08:21.246727 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:21Z","lastTransitionTime":"2025-12-03T00:08:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:21 crc kubenswrapper[4953]: I1203 00:08:21.347832 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xshk4" Dec 03 00:08:21 crc kubenswrapper[4953]: E1203 00:08:21.350189 4953 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xshk4" podUID="bd03a758-0ca9-47d8-86f7-18b2d610df43" Dec 03 00:08:21 crc kubenswrapper[4953]: I1203 00:08:21.351109 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:21 crc kubenswrapper[4953]: I1203 00:08:21.351232 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:21 crc kubenswrapper[4953]: I1203 00:08:21.351256 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:21 crc kubenswrapper[4953]: I1203 00:08:21.351285 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:21 crc kubenswrapper[4953]: I1203 00:08:21.351307 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:21Z","lastTransitionTime":"2025-12-03T00:08:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:21 crc kubenswrapper[4953]: I1203 00:08:21.366331 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7cf23d2d-21d4-493d-bdc7-d088aa876304\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4c95eb1f9b0c797376a1b557415b3f956890c971dc77c5257f238b4f32ff9b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ad2a6f5a3296c6a4f53f9da696e3e3f4d6c919f44b380c9b4114021b8ad74b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ad2a6f5a3296c6a4f53f9da696e3e3f4d6c919f44b380c9b4114021b8ad74b6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:06:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:06:51Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:08:21Z is after 2025-08-24T17:21:41Z" Dec 03 00:08:21 crc kubenswrapper[4953]: I1203 00:08:21.386260 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:08:21Z is after 2025-08-24T17:21:41Z" Dec 03 00:08:21 crc kubenswrapper[4953]: I1203 00:08:21.405735 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2ed9c7fa96c8951fbc1728fe7a91fefdd999077285621df9c1abedf9ae0c449\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:08:21Z is after 2025-08-24T17:21:41Z" Dec 03 00:08:21 crc kubenswrapper[4953]: I1203 00:08:21.428243 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:10Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:08:21Z is after 2025-08-24T17:21:41Z" Dec 03 00:08:21 crc kubenswrapper[4953]: I1203 00:08:21.452055 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-675qk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"43a4bad8-71f4-441a-8302-aeb526a986af\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:08:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:08:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e701c2ad05014683415ae2fb03b96f3acb38062ba7c5bfc4d0a9e34431695e36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5b5794fef6477aab5deed8f642e8cc50b37eb1fd0cc6f19c003f34ad31c1e3c0\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T00:08:02Z\\\",\\\"message\\\":\\\"2025-12-03T00:07:17+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_fa87bd16-684d-458f-b6b8-b6ada86b0d46\\\\n2025-12-03T00:07:17+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_fa87bd16-684d-458f-b6b8-b6ada86b0d46 to /host/opt/cni/bin/\\\\n2025-12-03T00:07:17Z [verbose] multus-daemon started\\\\n2025-12-03T00:07:17Z [verbose] Readiness Indicator file check\\\\n2025-12-03T00:08:02Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T00:07:16Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:08:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gdrw4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:07:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-675qk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:08:21Z is after 2025-08-24T17:21:41Z" Dec 03 00:08:21 crc kubenswrapper[4953]: I1203 00:08:21.454615 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:21 crc kubenswrapper[4953]: I1203 00:08:21.454688 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:21 crc kubenswrapper[4953]: I1203 00:08:21.454710 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:21 crc kubenswrapper[4953]: I1203 00:08:21.454739 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:21 crc kubenswrapper[4953]: I1203 00:08:21.454762 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:21Z","lastTransitionTime":"2025-12-03T00:08:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:21 crc kubenswrapper[4953]: I1203 00:08:21.473141 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f94f8c51-c665-4ca3-bf17-7a04ea5c7ade\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72c5cd7dc1c3a8adfedab6ba43a59d80b3efe65f2bb56be95e76cc86a409dcf3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2810ffc1b141fa7da811406053db9d72c708edf5f2bb1dd75e4d52ab35fd24c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c67b4277b20a9f1287f50ef82d0785c9d02543a8e7ed26e0413eae332b866db7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://258d78e1ad9d3dcda6976909ae3fe71aecead5b995fc66b3fd6e2722d80b5ab8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:06:51Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:08:21Z is after 2025-08-24T17:21:41Z" Dec 03 00:08:21 crc kubenswrapper[4953]: I1203 00:08:21.494896 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5652d0f3-3fd7-4af6-8941-9119e1177641\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2030aef5b7170991268fa5dc7e7ed8ce1fbe3a2765165a10c25251426575330f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ab6052f1abf17225155ce1bd86afa742e6e78f507b7deedd7161cd8fba9120b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://650b6c79cda2e1768ba1d8763db799e210ce2748893512068d25e60f821937ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d1d60162f7f6885d4f51b4ab44163cfae6d2d04355b1dd710e26b022450005df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d1d60162f7f6885d4f51b4ab44163cfae6d2d04355b1dd710e26b022450005df\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:06:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:06:51Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:08:21Z is after 2025-08-24T17:21:41Z" Dec 03 00:08:21 crc kubenswrapper[4953]: I1203 00:08:21.511484 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09657a9f448a4c92a759c73a0a892076d7eb46393e6fd565e5f73f34c71aac23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5af5eac4a58870f63497f99b503cfa5fb922278095e0332ec50376925ecf2109\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:08:21Z is after 2025-08-24T17:21:41Z" Dec 03 00:08:21 crc kubenswrapper[4953]: I1203 00:08:21.532264 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-wj6rz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"96aa67a0-15dd-420d-a1d5-ed239d7df1fb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01815cfd3f58c3f3f5f2a894c1790e0b69c5060c48dc0b1deff050b6ad7af916\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kf62m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b12bfb5bdf9fd4f2c50a335fc82df6ee0440eaf8741897da03aa2dae368cc6b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b12bfb5bdf9fd4f2c50a335fc82df6ee0440eaf8741897da03aa2dae368cc6b4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:07:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kf62m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83aaa5fa28110bc0ef05aff75cd1442bc9bbdd33b822ae2ef22d4e4ab83238bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://83aaa5fa28110bc0ef05aff75cd1442bc9bbdd33b822ae2ef22d4e4ab83238bc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:07:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:07:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kf62m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ab6213a33a19fedebce6d74a0ebbb3b3a2de5687504e21cc24093cc42ade635\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ab6213a33a19fedebce6d74a0ebbb3b3a2de5687504e21cc24093cc42ade635\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:07:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:07:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kf62m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://de480488843afa6b8dede1e1b42b5181dca1993e2f39dc1b461a18a57e3668e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://de480488843afa6b8dede1e1b42b5181dca1993e2f39dc1b461a18a57e3668e1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:07:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kf62m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://97be4ba692bc61078d4b0ba39fb0c12772548c6c9251ff3356ea702f151223cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://97be4ba692bc61078d4b0ba39fb0c12772548c6c9251ff3356ea702f151223cd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:07:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kf62m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1a1f009fcbd8af9217db1dc97f662719f8c1db232df0e2804e44c881abb1df2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b1a1f009fcbd8af9217db1dc97f662719f8c1db232df0e2804e44c881abb1df2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:07:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:07:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kf62m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:07:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-wj6rz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:08:21Z is after 2025-08-24T17:21:41Z" Dec 03 00:08:21 crc kubenswrapper[4953]: I1203 00:08:21.549328 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-kl5s9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d85e5c97-16ad-4bd7-9e73-b7f12678be97\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c21df91a9e4e3165686882b2eb559498224063dc633c3a63f958598108eca29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j24lb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:07:16Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-kl5s9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:08:21Z is after 2025-08-24T17:21:41Z" Dec 03 00:08:21 crc kubenswrapper[4953]: I1203 00:08:21.557956 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:21 crc kubenswrapper[4953]: I1203 00:08:21.558037 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:21 crc kubenswrapper[4953]: I1203 00:08:21.558064 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:21 crc kubenswrapper[4953]: I1203 00:08:21.558093 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:21 crc kubenswrapper[4953]: I1203 00:08:21.558309 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:21Z","lastTransitionTime":"2025-12-03T00:08:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:21 crc kubenswrapper[4953]: I1203 00:08:21.566186 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5cmlc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7e11a1d3-ecee-4bf0-87bf-085fab97b73d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d42390b9c53efc63add8ddcebf03c976f5e5448eaa2c34e076b2dc90dbac60fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6x2sc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:07:18Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5cmlc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:08:21Z is after 2025-08-24T17:21:41Z" Dec 03 00:08:21 crc kubenswrapper[4953]: I1203 00:08:21.583672 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-chmnz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a12fe6d-d25a-4b08-90da-cc9e510980a7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://711ab839cb9c0a90554dadbd45d810fc78aac09739340482fef8de8743b77283\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c982g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f802d5628a8cc32618044b9383211d5a492c56060d427f953b7793a1064029e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c982g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:07:29Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-chmnz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:08:21Z is after 2025-08-24T17:21:41Z" Dec 03 00:08:21 crc kubenswrapper[4953]: I1203 00:08:21.598188 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-xshk4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bd03a758-0ca9-47d8-86f7-18b2d610df43\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47gfh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47gfh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:07:31Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-xshk4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:08:21Z is after 2025-08-24T17:21:41Z" Dec 03 00:08:21 crc kubenswrapper[4953]: I1203 00:08:21.618776 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cf821840-5353-4620-8f73-99b56b371ab0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cae5e2994fcc24dc9b2ea73e79572615b3311edc42642b7f278fb84123989b23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://29e4b9d1b70a93fb133e0759ad148ae133766a66efeae3ff06a21dfae0e805ff\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dec43536772aa7c3f5a662fa7f3bd73b1cae2791ad454f85792287828d0ffeed\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://034e0bac6892d361700b8f33d6427b8a55e2bd3f89b14ea77a374c6be94d25e2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://761973393a93c59c23e69f41b8ac0d16abc35abd61ecbc6d51a42d94cd664233\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T00:07:09Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 00:07:03.713374 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 00:07:03.715052 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1528563048/tls.crt::/tmp/serving-cert-1528563048/tls.key\\\\\\\"\\\\nI1203 00:07:09.300052 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 00:07:09.304192 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 00:07:09.304255 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 00:07:09.304309 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 00:07:09.304323 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 00:07:09.313563 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 00:07:09.313597 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 00:07:09.313603 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 00:07:09.313609 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 00:07:09.313613 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 00:07:09.313618 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 00:07:09.313622 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 00:07:09.313709 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 00:07:09.327347 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T00:06:53Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c143d4232fb8bf7ef5999a664e431a00ab0db030f4d31a28d0bb21e00e38581\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:06:53Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72af6c27296eae2e275d455795c68f0d272869d63642b8353b17d466d0c901cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://72af6c27296eae2e275d455795c68f0d272869d63642b8353b17d466d0c901cf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:06:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:06:51Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:08:21Z is after 2025-08-24T17:21:41Z" Dec 03 00:08:21 crc kubenswrapper[4953]: I1203 00:08:21.638842 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:10Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:08:21Z is after 2025-08-24T17:21:41Z" Dec 03 00:08:21 crc kubenswrapper[4953]: I1203 00:08:21.658106 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce8b414e1fc22eb30decb6a6411b672e7f443055cb89ebb029fbf6ffc1d5dca7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:08:21Z is after 2025-08-24T17:21:41Z" Dec 03 00:08:21 crc kubenswrapper[4953]: I1203 00:08:21.661737 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:21 crc kubenswrapper[4953]: I1203 00:08:21.661800 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:21 crc kubenswrapper[4953]: I1203 00:08:21.661818 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:21 crc kubenswrapper[4953]: I1203 00:08:21.661843 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:21 crc kubenswrapper[4953]: I1203 00:08:21.661862 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:21Z","lastTransitionTime":"2025-12-03T00:08:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:21 crc kubenswrapper[4953]: I1203 00:08:21.673725 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-6dfbn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53ffb1e9-9a2e-4869-80d7-776f4aa7ef8b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ab4205c0d38ad6a0c32077eadb1fd5fd3a279d9c8b8a2ffbe693ec2c4018370c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ct26c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7df84405eda520ad5a22d35750edf07ef3c3977a756ad5739c54f91cba9da7b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ct26c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:07:16Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-6dfbn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:08:21Z is after 2025-08-24T17:21:41Z" Dec 03 00:08:21 crc kubenswrapper[4953]: I1203 00:08:21.705054 4953 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-txffx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8816ebee-c735-4b8d-b1ec-65349707b347\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:16Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:07:16Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4b85224419a4075acb340e4b11095f2adb5c316e285afa4e19295c2181c0d24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0f6f645a58993f052a6f365bb808f62d99f7fea74886c5cc022b2d86112321a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a61b951a9cbab2c48c0f2df8351f97d7ff71bf3674f641fe5eac30155232b224\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd5f428de229749485267e81d34e31b708728b52474032e88460e54747f21e11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://00a2ea4a9bb870fce30e7fa845f787f184e7a19e42c95511655ea8c5f84a26d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc0a5d4cc9d614d8492626a1b114bdc52f510eb5965d842c02c083f08d4849a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a4b625eda9b9daff8aa7fe3c5b94e417bcefe08fc6a65b4594f821cfdca7bfd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0a4b625eda9b9daff8aa7fe3c5b94e417bcefe08fc6a65b4594f821cfdca7bfd\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T00:08:13Z\\\",\\\"message\\\":\\\"il\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1203 00:08:13.521854 6989 transact.go:42] Configuring OVN: [{Op:update Table:Logical_Router_Static_Route Row:map[ip_prefix:10.217.0.0/22 nexthop:100.64.0.2 policy:{GoSet:[src-ip]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {8944024f-deb7-4076-afb3-4b50a2ff4b4b}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:mutate Table:Logical_Router Row:map[] Rows:[] Columns:[] Mutations:[{Column:static_routes Mutator:insert Value:{GoSet:[{GoUUID:8944024f-deb7-4076-afb3-4b50a2ff4b4b}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {f6d604c1-9711-4e25-be6c-79ec28bbad1b}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1203 00:08:13.522171 6989 obj_retry.go:551] Creating *factory.egressNode crc took: 1.771964ms\\\\nI1203 00:08:13.522207 6989 factory.go:1336] Added *v1.Node event handler 7\\\\nI1203 00:08:13.522242 6989 factory.go:1336] Added *v1.EgressIP event handler 8\\\\nI1203 00:08:13.522575 6989 factory.go:1336] Added *v1.EgressFirewall event handler 9\\\\nI1203 00:08:13.522717 6989 controller.go:132] Adding controller ef_node_controller event handlers\\\\nI1203 00:08:13.522757 6989 ovnkube.go:599] Stopped ovnkube\\\\nI1203 00:08:13.522794 6989 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1203 00:08:13.522871 6989 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T00:08:12Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-txffx_openshift-ovn-kubernetes(8816ebee-c735-4b8d-b1ec-65349707b347)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7deadc8e135ff1bc7c5a2f093e489bb5b4b6f166bc874c6acc2a841c5397f11e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af4c7764fca87144d01f3738ea010308744ccf50c54b7c471ccc213112de00d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af4c7764fca87144d01f3738ea010308744ccf50c54b7c471ccc213112de00d5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:07:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:07:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2fdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:07:16Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-txffx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:08:21Z is after 2025-08-24T17:21:41Z" Dec 03 00:08:21 crc kubenswrapper[4953]: I1203 00:08:21.764424 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:21 crc kubenswrapper[4953]: I1203 00:08:21.764476 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:21 crc kubenswrapper[4953]: I1203 00:08:21.764490 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:21 crc kubenswrapper[4953]: I1203 00:08:21.764509 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:21 crc kubenswrapper[4953]: I1203 00:08:21.764521 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:21Z","lastTransitionTime":"2025-12-03T00:08:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:21 crc kubenswrapper[4953]: I1203 00:08:21.867064 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:21 crc kubenswrapper[4953]: I1203 00:08:21.867119 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:21 crc kubenswrapper[4953]: I1203 00:08:21.867135 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:21 crc kubenswrapper[4953]: I1203 00:08:21.867159 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:21 crc kubenswrapper[4953]: I1203 00:08:21.867175 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:21Z","lastTransitionTime":"2025-12-03T00:08:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:21 crc kubenswrapper[4953]: I1203 00:08:21.969726 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:21 crc kubenswrapper[4953]: I1203 00:08:21.969817 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:21 crc kubenswrapper[4953]: I1203 00:08:21.969851 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:21 crc kubenswrapper[4953]: I1203 00:08:21.969882 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:21 crc kubenswrapper[4953]: I1203 00:08:21.969902 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:21Z","lastTransitionTime":"2025-12-03T00:08:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:22 crc kubenswrapper[4953]: I1203 00:08:22.072959 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:22 crc kubenswrapper[4953]: I1203 00:08:22.073043 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:22 crc kubenswrapper[4953]: I1203 00:08:22.073060 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:22 crc kubenswrapper[4953]: I1203 00:08:22.073083 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:22 crc kubenswrapper[4953]: I1203 00:08:22.073100 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:22Z","lastTransitionTime":"2025-12-03T00:08:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:22 crc kubenswrapper[4953]: I1203 00:08:22.176322 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:22 crc kubenswrapper[4953]: I1203 00:08:22.176383 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:22 crc kubenswrapper[4953]: I1203 00:08:22.176406 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:22 crc kubenswrapper[4953]: I1203 00:08:22.176433 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:22 crc kubenswrapper[4953]: I1203 00:08:22.176454 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:22Z","lastTransitionTime":"2025-12-03T00:08:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:22 crc kubenswrapper[4953]: I1203 00:08:22.279587 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:22 crc kubenswrapper[4953]: I1203 00:08:22.279647 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:22 crc kubenswrapper[4953]: I1203 00:08:22.279664 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:22 crc kubenswrapper[4953]: I1203 00:08:22.279686 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:22 crc kubenswrapper[4953]: I1203 00:08:22.279703 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:22Z","lastTransitionTime":"2025-12-03T00:08:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:22 crc kubenswrapper[4953]: I1203 00:08:22.348055 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 00:08:22 crc kubenswrapper[4953]: I1203 00:08:22.348098 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 00:08:22 crc kubenswrapper[4953]: I1203 00:08:22.348124 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 00:08:22 crc kubenswrapper[4953]: E1203 00:08:22.348242 4953 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 00:08:22 crc kubenswrapper[4953]: E1203 00:08:22.348363 4953 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 00:08:22 crc kubenswrapper[4953]: E1203 00:08:22.348502 4953 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 00:08:22 crc kubenswrapper[4953]: I1203 00:08:22.382832 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:22 crc kubenswrapper[4953]: I1203 00:08:22.382896 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:22 crc kubenswrapper[4953]: I1203 00:08:22.382919 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:22 crc kubenswrapper[4953]: I1203 00:08:22.382947 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:22 crc kubenswrapper[4953]: I1203 00:08:22.382970 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:22Z","lastTransitionTime":"2025-12-03T00:08:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:22 crc kubenswrapper[4953]: I1203 00:08:22.486370 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:22 crc kubenswrapper[4953]: I1203 00:08:22.486450 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:22 crc kubenswrapper[4953]: I1203 00:08:22.486478 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:22 crc kubenswrapper[4953]: I1203 00:08:22.486509 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:22 crc kubenswrapper[4953]: I1203 00:08:22.486532 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:22Z","lastTransitionTime":"2025-12-03T00:08:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:22 crc kubenswrapper[4953]: I1203 00:08:22.589770 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:22 crc kubenswrapper[4953]: I1203 00:08:22.589863 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:22 crc kubenswrapper[4953]: I1203 00:08:22.589881 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:22 crc kubenswrapper[4953]: I1203 00:08:22.589907 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:22 crc kubenswrapper[4953]: I1203 00:08:22.589925 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:22Z","lastTransitionTime":"2025-12-03T00:08:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:22 crc kubenswrapper[4953]: I1203 00:08:22.693154 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:22 crc kubenswrapper[4953]: I1203 00:08:22.693206 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:22 crc kubenswrapper[4953]: I1203 00:08:22.693225 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:22 crc kubenswrapper[4953]: I1203 00:08:22.693249 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:22 crc kubenswrapper[4953]: I1203 00:08:22.693267 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:22Z","lastTransitionTime":"2025-12-03T00:08:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:22 crc kubenswrapper[4953]: I1203 00:08:22.797234 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:22 crc kubenswrapper[4953]: I1203 00:08:22.797334 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:22 crc kubenswrapper[4953]: I1203 00:08:22.797355 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:22 crc kubenswrapper[4953]: I1203 00:08:22.797386 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:22 crc kubenswrapper[4953]: I1203 00:08:22.797407 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:22Z","lastTransitionTime":"2025-12-03T00:08:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:22 crc kubenswrapper[4953]: I1203 00:08:22.901057 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:22 crc kubenswrapper[4953]: I1203 00:08:22.901123 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:22 crc kubenswrapper[4953]: I1203 00:08:22.901140 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:22 crc kubenswrapper[4953]: I1203 00:08:22.901164 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:22 crc kubenswrapper[4953]: I1203 00:08:22.901181 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:22Z","lastTransitionTime":"2025-12-03T00:08:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:23 crc kubenswrapper[4953]: I1203 00:08:23.004900 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:23 crc kubenswrapper[4953]: I1203 00:08:23.005031 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:23 crc kubenswrapper[4953]: I1203 00:08:23.005072 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:23 crc kubenswrapper[4953]: I1203 00:08:23.005106 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:23 crc kubenswrapper[4953]: I1203 00:08:23.005131 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:23Z","lastTransitionTime":"2025-12-03T00:08:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:23 crc kubenswrapper[4953]: I1203 00:08:23.109773 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:23 crc kubenswrapper[4953]: I1203 00:08:23.109834 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:23 crc kubenswrapper[4953]: I1203 00:08:23.109852 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:23 crc kubenswrapper[4953]: I1203 00:08:23.109879 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:23 crc kubenswrapper[4953]: I1203 00:08:23.109899 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:23Z","lastTransitionTime":"2025-12-03T00:08:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:23 crc kubenswrapper[4953]: I1203 00:08:23.213311 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:23 crc kubenswrapper[4953]: I1203 00:08:23.213372 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:23 crc kubenswrapper[4953]: I1203 00:08:23.213395 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:23 crc kubenswrapper[4953]: I1203 00:08:23.213421 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:23 crc kubenswrapper[4953]: I1203 00:08:23.213443 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:23Z","lastTransitionTime":"2025-12-03T00:08:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:23 crc kubenswrapper[4953]: I1203 00:08:23.316081 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:23 crc kubenswrapper[4953]: I1203 00:08:23.316133 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:23 crc kubenswrapper[4953]: I1203 00:08:23.316156 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:23 crc kubenswrapper[4953]: I1203 00:08:23.316181 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:23 crc kubenswrapper[4953]: I1203 00:08:23.316202 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:23Z","lastTransitionTime":"2025-12-03T00:08:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:23 crc kubenswrapper[4953]: I1203 00:08:23.347473 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xshk4" Dec 03 00:08:23 crc kubenswrapper[4953]: E1203 00:08:23.347675 4953 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xshk4" podUID="bd03a758-0ca9-47d8-86f7-18b2d610df43" Dec 03 00:08:23 crc kubenswrapper[4953]: I1203 00:08:23.418475 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:23 crc kubenswrapper[4953]: I1203 00:08:23.418523 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:23 crc kubenswrapper[4953]: I1203 00:08:23.418534 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:23 crc kubenswrapper[4953]: I1203 00:08:23.418551 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:23 crc kubenswrapper[4953]: I1203 00:08:23.418562 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:23Z","lastTransitionTime":"2025-12-03T00:08:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:23 crc kubenswrapper[4953]: I1203 00:08:23.521346 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:23 crc kubenswrapper[4953]: I1203 00:08:23.521388 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:23 crc kubenswrapper[4953]: I1203 00:08:23.521399 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:23 crc kubenswrapper[4953]: I1203 00:08:23.521413 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:23 crc kubenswrapper[4953]: I1203 00:08:23.521424 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:23Z","lastTransitionTime":"2025-12-03T00:08:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:23 crc kubenswrapper[4953]: I1203 00:08:23.624355 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:23 crc kubenswrapper[4953]: I1203 00:08:23.624423 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:23 crc kubenswrapper[4953]: I1203 00:08:23.624435 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:23 crc kubenswrapper[4953]: I1203 00:08:23.624456 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:23 crc kubenswrapper[4953]: I1203 00:08:23.624470 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:23Z","lastTransitionTime":"2025-12-03T00:08:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:23 crc kubenswrapper[4953]: I1203 00:08:23.726876 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:23 crc kubenswrapper[4953]: I1203 00:08:23.726924 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:23 crc kubenswrapper[4953]: I1203 00:08:23.726935 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:23 crc kubenswrapper[4953]: I1203 00:08:23.726959 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:23 crc kubenswrapper[4953]: I1203 00:08:23.727003 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:23Z","lastTransitionTime":"2025-12-03T00:08:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:23 crc kubenswrapper[4953]: I1203 00:08:23.830284 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:23 crc kubenswrapper[4953]: I1203 00:08:23.830408 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:23 crc kubenswrapper[4953]: I1203 00:08:23.830429 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:23 crc kubenswrapper[4953]: I1203 00:08:23.830468 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:23 crc kubenswrapper[4953]: I1203 00:08:23.830486 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:23Z","lastTransitionTime":"2025-12-03T00:08:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:23 crc kubenswrapper[4953]: I1203 00:08:23.933197 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:23 crc kubenswrapper[4953]: I1203 00:08:23.933265 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:23 crc kubenswrapper[4953]: I1203 00:08:23.933287 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:23 crc kubenswrapper[4953]: I1203 00:08:23.933318 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:23 crc kubenswrapper[4953]: I1203 00:08:23.933340 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:23Z","lastTransitionTime":"2025-12-03T00:08:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:24 crc kubenswrapper[4953]: I1203 00:08:24.036030 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:24 crc kubenswrapper[4953]: I1203 00:08:24.036078 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:24 crc kubenswrapper[4953]: I1203 00:08:24.036093 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:24 crc kubenswrapper[4953]: I1203 00:08:24.036119 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:24 crc kubenswrapper[4953]: I1203 00:08:24.036135 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:24Z","lastTransitionTime":"2025-12-03T00:08:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:24 crc kubenswrapper[4953]: I1203 00:08:24.139293 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:24 crc kubenswrapper[4953]: I1203 00:08:24.139361 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:24 crc kubenswrapper[4953]: I1203 00:08:24.139381 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:24 crc kubenswrapper[4953]: I1203 00:08:24.139409 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:24 crc kubenswrapper[4953]: I1203 00:08:24.139430 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:24Z","lastTransitionTime":"2025-12-03T00:08:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:24 crc kubenswrapper[4953]: I1203 00:08:24.243201 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:24 crc kubenswrapper[4953]: I1203 00:08:24.243284 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:24 crc kubenswrapper[4953]: I1203 00:08:24.243304 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:24 crc kubenswrapper[4953]: I1203 00:08:24.243331 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:24 crc kubenswrapper[4953]: I1203 00:08:24.243359 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:24Z","lastTransitionTime":"2025-12-03T00:08:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:24 crc kubenswrapper[4953]: I1203 00:08:24.347066 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:24 crc kubenswrapper[4953]: I1203 00:08:24.347136 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:24 crc kubenswrapper[4953]: I1203 00:08:24.347161 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:24 crc kubenswrapper[4953]: I1203 00:08:24.347191 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:24 crc kubenswrapper[4953]: I1203 00:08:24.347216 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:24Z","lastTransitionTime":"2025-12-03T00:08:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:24 crc kubenswrapper[4953]: I1203 00:08:24.347273 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 00:08:24 crc kubenswrapper[4953]: I1203 00:08:24.347341 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 00:08:24 crc kubenswrapper[4953]: E1203 00:08:24.347422 4953 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 00:08:24 crc kubenswrapper[4953]: E1203 00:08:24.347520 4953 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 00:08:24 crc kubenswrapper[4953]: I1203 00:08:24.347674 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 00:08:24 crc kubenswrapper[4953]: E1203 00:08:24.347938 4953 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 00:08:24 crc kubenswrapper[4953]: I1203 00:08:24.450236 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:24 crc kubenswrapper[4953]: I1203 00:08:24.450280 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:24 crc kubenswrapper[4953]: I1203 00:08:24.450290 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:24 crc kubenswrapper[4953]: I1203 00:08:24.450307 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:24 crc kubenswrapper[4953]: I1203 00:08:24.450319 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:24Z","lastTransitionTime":"2025-12-03T00:08:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:24 crc kubenswrapper[4953]: I1203 00:08:24.553732 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:24 crc kubenswrapper[4953]: I1203 00:08:24.553794 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:24 crc kubenswrapper[4953]: I1203 00:08:24.553816 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:24 crc kubenswrapper[4953]: I1203 00:08:24.553857 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:24 crc kubenswrapper[4953]: I1203 00:08:24.553876 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:24Z","lastTransitionTime":"2025-12-03T00:08:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:24 crc kubenswrapper[4953]: I1203 00:08:24.657264 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:24 crc kubenswrapper[4953]: I1203 00:08:24.657334 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:24 crc kubenswrapper[4953]: I1203 00:08:24.657351 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:24 crc kubenswrapper[4953]: I1203 00:08:24.657377 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:24 crc kubenswrapper[4953]: I1203 00:08:24.657396 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:24Z","lastTransitionTime":"2025-12-03T00:08:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:24 crc kubenswrapper[4953]: I1203 00:08:24.760275 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:24 crc kubenswrapper[4953]: I1203 00:08:24.760333 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:24 crc kubenswrapper[4953]: I1203 00:08:24.760349 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:24 crc kubenswrapper[4953]: I1203 00:08:24.760372 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:24 crc kubenswrapper[4953]: I1203 00:08:24.760391 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:24Z","lastTransitionTime":"2025-12-03T00:08:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:24 crc kubenswrapper[4953]: I1203 00:08:24.842474 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:24 crc kubenswrapper[4953]: I1203 00:08:24.842547 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:24 crc kubenswrapper[4953]: I1203 00:08:24.842565 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:24 crc kubenswrapper[4953]: I1203 00:08:24.842589 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:24 crc kubenswrapper[4953]: I1203 00:08:24.842608 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:24Z","lastTransitionTime":"2025-12-03T00:08:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:24 crc kubenswrapper[4953]: E1203 00:08:24.863946 4953 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:08:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:08:24Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:08:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:08:24Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:08:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:08:24Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:08:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:08:24Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"eb4ae5a6-0f6f-49ef-a027-1cdee9d3c536\\\",\\\"systemUUID\\\":\\\"4c1f3d44-3d3d-425d-993f-471223206a89\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:08:24Z is after 2025-08-24T17:21:41Z" Dec 03 00:08:24 crc kubenswrapper[4953]: I1203 00:08:24.870442 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:24 crc kubenswrapper[4953]: I1203 00:08:24.870525 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:24 crc kubenswrapper[4953]: I1203 00:08:24.870549 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:24 crc kubenswrapper[4953]: I1203 00:08:24.870581 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:24 crc kubenswrapper[4953]: I1203 00:08:24.870622 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:24Z","lastTransitionTime":"2025-12-03T00:08:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:24 crc kubenswrapper[4953]: E1203 00:08:24.894023 4953 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:08:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:08:24Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:08:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:08:24Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:08:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:08:24Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:08:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:08:24Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"eb4ae5a6-0f6f-49ef-a027-1cdee9d3c536\\\",\\\"systemUUID\\\":\\\"4c1f3d44-3d3d-425d-993f-471223206a89\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:08:24Z is after 2025-08-24T17:21:41Z" Dec 03 00:08:24 crc kubenswrapper[4953]: I1203 00:08:24.899796 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:24 crc kubenswrapper[4953]: I1203 00:08:24.899888 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:24 crc kubenswrapper[4953]: I1203 00:08:24.899906 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:24 crc kubenswrapper[4953]: I1203 00:08:24.899929 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:24 crc kubenswrapper[4953]: I1203 00:08:24.899947 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:24Z","lastTransitionTime":"2025-12-03T00:08:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:24 crc kubenswrapper[4953]: E1203 00:08:24.921927 4953 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:08:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:08:24Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:08:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:08:24Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:08:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:08:24Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:08:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:08:24Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"eb4ae5a6-0f6f-49ef-a027-1cdee9d3c536\\\",\\\"systemUUID\\\":\\\"4c1f3d44-3d3d-425d-993f-471223206a89\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:08:24Z is after 2025-08-24T17:21:41Z" Dec 03 00:08:24 crc kubenswrapper[4953]: I1203 00:08:24.927618 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:24 crc kubenswrapper[4953]: I1203 00:08:24.927720 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:24 crc kubenswrapper[4953]: I1203 00:08:24.927741 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:24 crc kubenswrapper[4953]: I1203 00:08:24.927808 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:24 crc kubenswrapper[4953]: I1203 00:08:24.927835 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:24Z","lastTransitionTime":"2025-12-03T00:08:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:24 crc kubenswrapper[4953]: E1203 00:08:24.951087 4953 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:08:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:08:24Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:08:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:08:24Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:08:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:08:24Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:08:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:08:24Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"eb4ae5a6-0f6f-49ef-a027-1cdee9d3c536\\\",\\\"systemUUID\\\":\\\"4c1f3d44-3d3d-425d-993f-471223206a89\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:08:24Z is after 2025-08-24T17:21:41Z" Dec 03 00:08:24 crc kubenswrapper[4953]: I1203 00:08:24.956382 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:24 crc kubenswrapper[4953]: I1203 00:08:24.956440 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:24 crc kubenswrapper[4953]: I1203 00:08:24.956457 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:24 crc kubenswrapper[4953]: I1203 00:08:24.956481 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:24 crc kubenswrapper[4953]: I1203 00:08:24.956499 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:24Z","lastTransitionTime":"2025-12-03T00:08:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:24 crc kubenswrapper[4953]: E1203 00:08:24.976243 4953 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:08:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:08:24Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:08:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:08:24Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:08:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:08:24Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:08:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:08:24Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"eb4ae5a6-0f6f-49ef-a027-1cdee9d3c536\\\",\\\"systemUUID\\\":\\\"4c1f3d44-3d3d-425d-993f-471223206a89\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:08:24Z is after 2025-08-24T17:21:41Z" Dec 03 00:08:24 crc kubenswrapper[4953]: E1203 00:08:24.976478 4953 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 03 00:08:24 crc kubenswrapper[4953]: I1203 00:08:24.978713 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:24 crc kubenswrapper[4953]: I1203 00:08:24.978805 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:24 crc kubenswrapper[4953]: I1203 00:08:24.978834 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:24 crc kubenswrapper[4953]: I1203 00:08:24.978867 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:24 crc kubenswrapper[4953]: I1203 00:08:24.978892 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:24Z","lastTransitionTime":"2025-12-03T00:08:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:25 crc kubenswrapper[4953]: I1203 00:08:25.081815 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:25 crc kubenswrapper[4953]: I1203 00:08:25.081872 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:25 crc kubenswrapper[4953]: I1203 00:08:25.081889 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:25 crc kubenswrapper[4953]: I1203 00:08:25.081915 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:25 crc kubenswrapper[4953]: I1203 00:08:25.081932 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:25Z","lastTransitionTime":"2025-12-03T00:08:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:25 crc kubenswrapper[4953]: I1203 00:08:25.189885 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:25 crc kubenswrapper[4953]: I1203 00:08:25.189953 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:25 crc kubenswrapper[4953]: I1203 00:08:25.189987 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:25 crc kubenswrapper[4953]: I1203 00:08:25.190011 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:25 crc kubenswrapper[4953]: I1203 00:08:25.190029 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:25Z","lastTransitionTime":"2025-12-03T00:08:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:25 crc kubenswrapper[4953]: I1203 00:08:25.292637 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:25 crc kubenswrapper[4953]: I1203 00:08:25.292697 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:25 crc kubenswrapper[4953]: I1203 00:08:25.292713 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:25 crc kubenswrapper[4953]: I1203 00:08:25.292737 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:25 crc kubenswrapper[4953]: I1203 00:08:25.292756 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:25Z","lastTransitionTime":"2025-12-03T00:08:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:25 crc kubenswrapper[4953]: I1203 00:08:25.347677 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xshk4" Dec 03 00:08:25 crc kubenswrapper[4953]: E1203 00:08:25.347880 4953 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xshk4" podUID="bd03a758-0ca9-47d8-86f7-18b2d610df43" Dec 03 00:08:25 crc kubenswrapper[4953]: I1203 00:08:25.395470 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:25 crc kubenswrapper[4953]: I1203 00:08:25.395524 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:25 crc kubenswrapper[4953]: I1203 00:08:25.395542 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:25 crc kubenswrapper[4953]: I1203 00:08:25.395565 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:25 crc kubenswrapper[4953]: I1203 00:08:25.395586 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:25Z","lastTransitionTime":"2025-12-03T00:08:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:25 crc kubenswrapper[4953]: I1203 00:08:25.498608 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:25 crc kubenswrapper[4953]: I1203 00:08:25.498668 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:25 crc kubenswrapper[4953]: I1203 00:08:25.498685 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:25 crc kubenswrapper[4953]: I1203 00:08:25.498708 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:25 crc kubenswrapper[4953]: I1203 00:08:25.498725 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:25Z","lastTransitionTime":"2025-12-03T00:08:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:25 crc kubenswrapper[4953]: I1203 00:08:25.601538 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:25 crc kubenswrapper[4953]: I1203 00:08:25.601607 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:25 crc kubenswrapper[4953]: I1203 00:08:25.601630 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:25 crc kubenswrapper[4953]: I1203 00:08:25.601659 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:25 crc kubenswrapper[4953]: I1203 00:08:25.601679 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:25Z","lastTransitionTime":"2025-12-03T00:08:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:25 crc kubenswrapper[4953]: I1203 00:08:25.705477 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:25 crc kubenswrapper[4953]: I1203 00:08:25.705541 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:25 crc kubenswrapper[4953]: I1203 00:08:25.705562 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:25 crc kubenswrapper[4953]: I1203 00:08:25.705591 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:25 crc kubenswrapper[4953]: I1203 00:08:25.705611 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:25Z","lastTransitionTime":"2025-12-03T00:08:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:25 crc kubenswrapper[4953]: I1203 00:08:25.808923 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:25 crc kubenswrapper[4953]: I1203 00:08:25.809062 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:25 crc kubenswrapper[4953]: I1203 00:08:25.809088 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:25 crc kubenswrapper[4953]: I1203 00:08:25.809124 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:25 crc kubenswrapper[4953]: I1203 00:08:25.809147 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:25Z","lastTransitionTime":"2025-12-03T00:08:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:25 crc kubenswrapper[4953]: I1203 00:08:25.911789 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:25 crc kubenswrapper[4953]: I1203 00:08:25.911940 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:25 crc kubenswrapper[4953]: I1203 00:08:25.911966 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:25 crc kubenswrapper[4953]: I1203 00:08:25.912049 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:25 crc kubenswrapper[4953]: I1203 00:08:25.912072 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:25Z","lastTransitionTime":"2025-12-03T00:08:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:26 crc kubenswrapper[4953]: I1203 00:08:26.014894 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:26 crc kubenswrapper[4953]: I1203 00:08:26.014948 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:26 crc kubenswrapper[4953]: I1203 00:08:26.014966 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:26 crc kubenswrapper[4953]: I1203 00:08:26.015020 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:26 crc kubenswrapper[4953]: I1203 00:08:26.015042 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:26Z","lastTransitionTime":"2025-12-03T00:08:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:26 crc kubenswrapper[4953]: I1203 00:08:26.118234 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:26 crc kubenswrapper[4953]: I1203 00:08:26.118296 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:26 crc kubenswrapper[4953]: I1203 00:08:26.118312 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:26 crc kubenswrapper[4953]: I1203 00:08:26.118334 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:26 crc kubenswrapper[4953]: I1203 00:08:26.118350 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:26Z","lastTransitionTime":"2025-12-03T00:08:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:26 crc kubenswrapper[4953]: I1203 00:08:26.220941 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:26 crc kubenswrapper[4953]: I1203 00:08:26.221025 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:26 crc kubenswrapper[4953]: I1203 00:08:26.221049 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:26 crc kubenswrapper[4953]: I1203 00:08:26.221078 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:26 crc kubenswrapper[4953]: I1203 00:08:26.221099 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:26Z","lastTransitionTime":"2025-12-03T00:08:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:26 crc kubenswrapper[4953]: I1203 00:08:26.324461 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:26 crc kubenswrapper[4953]: I1203 00:08:26.324513 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:26 crc kubenswrapper[4953]: I1203 00:08:26.324530 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:26 crc kubenswrapper[4953]: I1203 00:08:26.324555 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:26 crc kubenswrapper[4953]: I1203 00:08:26.324573 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:26Z","lastTransitionTime":"2025-12-03T00:08:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:26 crc kubenswrapper[4953]: I1203 00:08:26.347171 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 00:08:26 crc kubenswrapper[4953]: I1203 00:08:26.347253 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 00:08:26 crc kubenswrapper[4953]: I1203 00:08:26.347207 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 00:08:26 crc kubenswrapper[4953]: E1203 00:08:26.347451 4953 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 00:08:26 crc kubenswrapper[4953]: E1203 00:08:26.347536 4953 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 00:08:26 crc kubenswrapper[4953]: E1203 00:08:26.347740 4953 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 00:08:26 crc kubenswrapper[4953]: I1203 00:08:26.427041 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:26 crc kubenswrapper[4953]: I1203 00:08:26.427082 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:26 crc kubenswrapper[4953]: I1203 00:08:26.427090 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:26 crc kubenswrapper[4953]: I1203 00:08:26.427106 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:26 crc kubenswrapper[4953]: I1203 00:08:26.427115 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:26Z","lastTransitionTime":"2025-12-03T00:08:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:26 crc kubenswrapper[4953]: I1203 00:08:26.530312 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:26 crc kubenswrapper[4953]: I1203 00:08:26.530351 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:26 crc kubenswrapper[4953]: I1203 00:08:26.530361 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:26 crc kubenswrapper[4953]: I1203 00:08:26.530376 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:26 crc kubenswrapper[4953]: I1203 00:08:26.530385 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:26Z","lastTransitionTime":"2025-12-03T00:08:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:26 crc kubenswrapper[4953]: I1203 00:08:26.633215 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:26 crc kubenswrapper[4953]: I1203 00:08:26.633314 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:26 crc kubenswrapper[4953]: I1203 00:08:26.633334 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:26 crc kubenswrapper[4953]: I1203 00:08:26.633412 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:26 crc kubenswrapper[4953]: I1203 00:08:26.633437 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:26Z","lastTransitionTime":"2025-12-03T00:08:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:26 crc kubenswrapper[4953]: I1203 00:08:26.737314 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:26 crc kubenswrapper[4953]: I1203 00:08:26.737384 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:26 crc kubenswrapper[4953]: I1203 00:08:26.737400 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:26 crc kubenswrapper[4953]: I1203 00:08:26.737422 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:26 crc kubenswrapper[4953]: I1203 00:08:26.737436 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:26Z","lastTransitionTime":"2025-12-03T00:08:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:26 crc kubenswrapper[4953]: I1203 00:08:26.840586 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:26 crc kubenswrapper[4953]: I1203 00:08:26.840660 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:26 crc kubenswrapper[4953]: I1203 00:08:26.840681 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:26 crc kubenswrapper[4953]: I1203 00:08:26.840708 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:26 crc kubenswrapper[4953]: I1203 00:08:26.840748 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:26Z","lastTransitionTime":"2025-12-03T00:08:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:26 crc kubenswrapper[4953]: I1203 00:08:26.943252 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:26 crc kubenswrapper[4953]: I1203 00:08:26.943316 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:26 crc kubenswrapper[4953]: I1203 00:08:26.943335 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:26 crc kubenswrapper[4953]: I1203 00:08:26.943358 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:26 crc kubenswrapper[4953]: I1203 00:08:26.943375 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:26Z","lastTransitionTime":"2025-12-03T00:08:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:27 crc kubenswrapper[4953]: I1203 00:08:27.046365 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:27 crc kubenswrapper[4953]: I1203 00:08:27.046419 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:27 crc kubenswrapper[4953]: I1203 00:08:27.046437 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:27 crc kubenswrapper[4953]: I1203 00:08:27.046459 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:27 crc kubenswrapper[4953]: I1203 00:08:27.046501 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:27Z","lastTransitionTime":"2025-12-03T00:08:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:27 crc kubenswrapper[4953]: I1203 00:08:27.148950 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:27 crc kubenswrapper[4953]: I1203 00:08:27.149045 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:27 crc kubenswrapper[4953]: I1203 00:08:27.149064 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:27 crc kubenswrapper[4953]: I1203 00:08:27.149088 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:27 crc kubenswrapper[4953]: I1203 00:08:27.149106 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:27Z","lastTransitionTime":"2025-12-03T00:08:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:27 crc kubenswrapper[4953]: I1203 00:08:27.253124 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:27 crc kubenswrapper[4953]: I1203 00:08:27.253199 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:27 crc kubenswrapper[4953]: I1203 00:08:27.253219 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:27 crc kubenswrapper[4953]: I1203 00:08:27.253299 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:27 crc kubenswrapper[4953]: I1203 00:08:27.253330 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:27Z","lastTransitionTime":"2025-12-03T00:08:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:27 crc kubenswrapper[4953]: I1203 00:08:27.347963 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xshk4" Dec 03 00:08:27 crc kubenswrapper[4953]: E1203 00:08:27.348658 4953 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xshk4" podUID="bd03a758-0ca9-47d8-86f7-18b2d610df43" Dec 03 00:08:27 crc kubenswrapper[4953]: I1203 00:08:27.349364 4953 scope.go:117] "RemoveContainer" containerID="0a4b625eda9b9daff8aa7fe3c5b94e417bcefe08fc6a65b4594f821cfdca7bfd" Dec 03 00:08:27 crc kubenswrapper[4953]: E1203 00:08:27.349674 4953 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-txffx_openshift-ovn-kubernetes(8816ebee-c735-4b8d-b1ec-65349707b347)\"" pod="openshift-ovn-kubernetes/ovnkube-node-txffx" podUID="8816ebee-c735-4b8d-b1ec-65349707b347" Dec 03 00:08:27 crc kubenswrapper[4953]: I1203 00:08:27.355493 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:27 crc kubenswrapper[4953]: I1203 00:08:27.355537 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:27 crc kubenswrapper[4953]: I1203 00:08:27.355555 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:27 crc kubenswrapper[4953]: I1203 00:08:27.355577 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:27 crc kubenswrapper[4953]: I1203 00:08:27.355594 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:27Z","lastTransitionTime":"2025-12-03T00:08:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:27 crc kubenswrapper[4953]: I1203 00:08:27.459143 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:27 crc kubenswrapper[4953]: I1203 00:08:27.459199 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:27 crc kubenswrapper[4953]: I1203 00:08:27.459215 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:27 crc kubenswrapper[4953]: I1203 00:08:27.459238 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:27 crc kubenswrapper[4953]: I1203 00:08:27.459259 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:27Z","lastTransitionTime":"2025-12-03T00:08:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:27 crc kubenswrapper[4953]: I1203 00:08:27.562413 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:27 crc kubenswrapper[4953]: I1203 00:08:27.562479 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:27 crc kubenswrapper[4953]: I1203 00:08:27.562496 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:27 crc kubenswrapper[4953]: I1203 00:08:27.562527 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:27 crc kubenswrapper[4953]: I1203 00:08:27.562549 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:27Z","lastTransitionTime":"2025-12-03T00:08:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:27 crc kubenswrapper[4953]: I1203 00:08:27.666445 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:27 crc kubenswrapper[4953]: I1203 00:08:27.666516 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:27 crc kubenswrapper[4953]: I1203 00:08:27.666536 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:27 crc kubenswrapper[4953]: I1203 00:08:27.666560 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:27 crc kubenswrapper[4953]: I1203 00:08:27.666577 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:27Z","lastTransitionTime":"2025-12-03T00:08:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:27 crc kubenswrapper[4953]: I1203 00:08:27.769041 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:27 crc kubenswrapper[4953]: I1203 00:08:27.769112 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:27 crc kubenswrapper[4953]: I1203 00:08:27.769133 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:27 crc kubenswrapper[4953]: I1203 00:08:27.769158 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:27 crc kubenswrapper[4953]: I1203 00:08:27.769175 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:27Z","lastTransitionTime":"2025-12-03T00:08:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:27 crc kubenswrapper[4953]: I1203 00:08:27.871809 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:27 crc kubenswrapper[4953]: I1203 00:08:27.871847 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:27 crc kubenswrapper[4953]: I1203 00:08:27.871857 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:27 crc kubenswrapper[4953]: I1203 00:08:27.871877 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:27 crc kubenswrapper[4953]: I1203 00:08:27.871892 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:27Z","lastTransitionTime":"2025-12-03T00:08:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:27 crc kubenswrapper[4953]: I1203 00:08:27.974780 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:27 crc kubenswrapper[4953]: I1203 00:08:27.974830 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:27 crc kubenswrapper[4953]: I1203 00:08:27.974847 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:27 crc kubenswrapper[4953]: I1203 00:08:27.974875 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:27 crc kubenswrapper[4953]: I1203 00:08:27.974893 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:27Z","lastTransitionTime":"2025-12-03T00:08:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:28 crc kubenswrapper[4953]: I1203 00:08:28.077280 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:28 crc kubenswrapper[4953]: I1203 00:08:28.077360 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:28 crc kubenswrapper[4953]: I1203 00:08:28.077378 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:28 crc kubenswrapper[4953]: I1203 00:08:28.077404 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:28 crc kubenswrapper[4953]: I1203 00:08:28.077425 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:28Z","lastTransitionTime":"2025-12-03T00:08:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:28 crc kubenswrapper[4953]: I1203 00:08:28.180343 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:28 crc kubenswrapper[4953]: I1203 00:08:28.180414 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:28 crc kubenswrapper[4953]: I1203 00:08:28.180432 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:28 crc kubenswrapper[4953]: I1203 00:08:28.180456 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:28 crc kubenswrapper[4953]: I1203 00:08:28.180476 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:28Z","lastTransitionTime":"2025-12-03T00:08:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:28 crc kubenswrapper[4953]: I1203 00:08:28.283501 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:28 crc kubenswrapper[4953]: I1203 00:08:28.283543 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:28 crc kubenswrapper[4953]: I1203 00:08:28.283556 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:28 crc kubenswrapper[4953]: I1203 00:08:28.283573 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:28 crc kubenswrapper[4953]: I1203 00:08:28.283587 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:28Z","lastTransitionTime":"2025-12-03T00:08:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:28 crc kubenswrapper[4953]: I1203 00:08:28.347143 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 00:08:28 crc kubenswrapper[4953]: I1203 00:08:28.347183 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 00:08:28 crc kubenswrapper[4953]: I1203 00:08:28.347143 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 00:08:28 crc kubenswrapper[4953]: E1203 00:08:28.347304 4953 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 00:08:28 crc kubenswrapper[4953]: E1203 00:08:28.347436 4953 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 00:08:28 crc kubenswrapper[4953]: E1203 00:08:28.347767 4953 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 00:08:28 crc kubenswrapper[4953]: I1203 00:08:28.386074 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:28 crc kubenswrapper[4953]: I1203 00:08:28.386148 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:28 crc kubenswrapper[4953]: I1203 00:08:28.386177 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:28 crc kubenswrapper[4953]: I1203 00:08:28.386206 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:28 crc kubenswrapper[4953]: I1203 00:08:28.386227 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:28Z","lastTransitionTime":"2025-12-03T00:08:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:28 crc kubenswrapper[4953]: I1203 00:08:28.488934 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:28 crc kubenswrapper[4953]: I1203 00:08:28.489047 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:28 crc kubenswrapper[4953]: I1203 00:08:28.489076 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:28 crc kubenswrapper[4953]: I1203 00:08:28.489103 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:28 crc kubenswrapper[4953]: I1203 00:08:28.489124 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:28Z","lastTransitionTime":"2025-12-03T00:08:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:28 crc kubenswrapper[4953]: I1203 00:08:28.592728 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:28 crc kubenswrapper[4953]: I1203 00:08:28.592823 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:28 crc kubenswrapper[4953]: I1203 00:08:28.592844 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:28 crc kubenswrapper[4953]: I1203 00:08:28.592866 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:28 crc kubenswrapper[4953]: I1203 00:08:28.592884 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:28Z","lastTransitionTime":"2025-12-03T00:08:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:28 crc kubenswrapper[4953]: I1203 00:08:28.702281 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:28 crc kubenswrapper[4953]: I1203 00:08:28.702347 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:28 crc kubenswrapper[4953]: I1203 00:08:28.702364 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:28 crc kubenswrapper[4953]: I1203 00:08:28.702390 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:28 crc kubenswrapper[4953]: I1203 00:08:28.702408 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:28Z","lastTransitionTime":"2025-12-03T00:08:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:28 crc kubenswrapper[4953]: I1203 00:08:28.806115 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:28 crc kubenswrapper[4953]: I1203 00:08:28.806539 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:28 crc kubenswrapper[4953]: I1203 00:08:28.806761 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:28 crc kubenswrapper[4953]: I1203 00:08:28.807030 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:28 crc kubenswrapper[4953]: I1203 00:08:28.807249 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:28Z","lastTransitionTime":"2025-12-03T00:08:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:28 crc kubenswrapper[4953]: I1203 00:08:28.910060 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:28 crc kubenswrapper[4953]: I1203 00:08:28.910132 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:28 crc kubenswrapper[4953]: I1203 00:08:28.910150 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:28 crc kubenswrapper[4953]: I1203 00:08:28.910177 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:28 crc kubenswrapper[4953]: I1203 00:08:28.910196 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:28Z","lastTransitionTime":"2025-12-03T00:08:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:29 crc kubenswrapper[4953]: I1203 00:08:29.013150 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:29 crc kubenswrapper[4953]: I1203 00:08:29.013536 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:29 crc kubenswrapper[4953]: I1203 00:08:29.013745 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:29 crc kubenswrapper[4953]: I1203 00:08:29.013935 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:29 crc kubenswrapper[4953]: I1203 00:08:29.014154 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:29Z","lastTransitionTime":"2025-12-03T00:08:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:29 crc kubenswrapper[4953]: I1203 00:08:29.116619 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:29 crc kubenswrapper[4953]: I1203 00:08:29.117024 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:29 crc kubenswrapper[4953]: I1203 00:08:29.117206 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:29 crc kubenswrapper[4953]: I1203 00:08:29.117388 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:29 crc kubenswrapper[4953]: I1203 00:08:29.117553 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:29Z","lastTransitionTime":"2025-12-03T00:08:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:29 crc kubenswrapper[4953]: I1203 00:08:29.221333 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:29 crc kubenswrapper[4953]: I1203 00:08:29.221380 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:29 crc kubenswrapper[4953]: I1203 00:08:29.221399 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:29 crc kubenswrapper[4953]: I1203 00:08:29.221422 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:29 crc kubenswrapper[4953]: I1203 00:08:29.221439 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:29Z","lastTransitionTime":"2025-12-03T00:08:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:29 crc kubenswrapper[4953]: I1203 00:08:29.324038 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:29 crc kubenswrapper[4953]: I1203 00:08:29.324163 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:29 crc kubenswrapper[4953]: I1203 00:08:29.324189 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:29 crc kubenswrapper[4953]: I1203 00:08:29.324220 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:29 crc kubenswrapper[4953]: I1203 00:08:29.324242 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:29Z","lastTransitionTime":"2025-12-03T00:08:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:29 crc kubenswrapper[4953]: I1203 00:08:29.347474 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xshk4" Dec 03 00:08:29 crc kubenswrapper[4953]: E1203 00:08:29.347649 4953 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xshk4" podUID="bd03a758-0ca9-47d8-86f7-18b2d610df43" Dec 03 00:08:29 crc kubenswrapper[4953]: I1203 00:08:29.427327 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:29 crc kubenswrapper[4953]: I1203 00:08:29.427415 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:29 crc kubenswrapper[4953]: I1203 00:08:29.427448 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:29 crc kubenswrapper[4953]: I1203 00:08:29.427477 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:29 crc kubenswrapper[4953]: I1203 00:08:29.427497 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:29Z","lastTransitionTime":"2025-12-03T00:08:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:29 crc kubenswrapper[4953]: I1203 00:08:29.530809 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:29 crc kubenswrapper[4953]: I1203 00:08:29.530937 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:29 crc kubenswrapper[4953]: I1203 00:08:29.530950 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:29 crc kubenswrapper[4953]: I1203 00:08:29.530967 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:29 crc kubenswrapper[4953]: I1203 00:08:29.530999 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:29Z","lastTransitionTime":"2025-12-03T00:08:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:29 crc kubenswrapper[4953]: I1203 00:08:29.634042 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:29 crc kubenswrapper[4953]: I1203 00:08:29.634101 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:29 crc kubenswrapper[4953]: I1203 00:08:29.634117 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:29 crc kubenswrapper[4953]: I1203 00:08:29.634142 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:29 crc kubenswrapper[4953]: I1203 00:08:29.634162 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:29Z","lastTransitionTime":"2025-12-03T00:08:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:29 crc kubenswrapper[4953]: I1203 00:08:29.737545 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:29 crc kubenswrapper[4953]: I1203 00:08:29.737602 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:29 crc kubenswrapper[4953]: I1203 00:08:29.737624 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:29 crc kubenswrapper[4953]: I1203 00:08:29.737655 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:29 crc kubenswrapper[4953]: I1203 00:08:29.737677 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:29Z","lastTransitionTime":"2025-12-03T00:08:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:29 crc kubenswrapper[4953]: I1203 00:08:29.840483 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:29 crc kubenswrapper[4953]: I1203 00:08:29.840535 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:29 crc kubenswrapper[4953]: I1203 00:08:29.840552 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:29 crc kubenswrapper[4953]: I1203 00:08:29.840574 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:29 crc kubenswrapper[4953]: I1203 00:08:29.840590 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:29Z","lastTransitionTime":"2025-12-03T00:08:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:29 crc kubenswrapper[4953]: I1203 00:08:29.943806 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:29 crc kubenswrapper[4953]: I1203 00:08:29.943837 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:29 crc kubenswrapper[4953]: I1203 00:08:29.943845 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:29 crc kubenswrapper[4953]: I1203 00:08:29.943857 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:29 crc kubenswrapper[4953]: I1203 00:08:29.943874 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:29Z","lastTransitionTime":"2025-12-03T00:08:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:30 crc kubenswrapper[4953]: I1203 00:08:30.046086 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:30 crc kubenswrapper[4953]: I1203 00:08:30.046131 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:30 crc kubenswrapper[4953]: I1203 00:08:30.046143 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:30 crc kubenswrapper[4953]: I1203 00:08:30.046159 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:30 crc kubenswrapper[4953]: I1203 00:08:30.046172 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:30Z","lastTransitionTime":"2025-12-03T00:08:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:30 crc kubenswrapper[4953]: I1203 00:08:30.149089 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:30 crc kubenswrapper[4953]: I1203 00:08:30.149155 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:30 crc kubenswrapper[4953]: I1203 00:08:30.149175 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:30 crc kubenswrapper[4953]: I1203 00:08:30.149199 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:30 crc kubenswrapper[4953]: I1203 00:08:30.149219 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:30Z","lastTransitionTime":"2025-12-03T00:08:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:30 crc kubenswrapper[4953]: I1203 00:08:30.251910 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:30 crc kubenswrapper[4953]: I1203 00:08:30.252012 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:30 crc kubenswrapper[4953]: I1203 00:08:30.252048 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:30 crc kubenswrapper[4953]: I1203 00:08:30.252108 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:30 crc kubenswrapper[4953]: I1203 00:08:30.252132 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:30Z","lastTransitionTime":"2025-12-03T00:08:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:30 crc kubenswrapper[4953]: I1203 00:08:30.347602 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 00:08:30 crc kubenswrapper[4953]: E1203 00:08:30.347801 4953 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 00:08:30 crc kubenswrapper[4953]: I1203 00:08:30.348190 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 00:08:30 crc kubenswrapper[4953]: E1203 00:08:30.348332 4953 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 00:08:30 crc kubenswrapper[4953]: I1203 00:08:30.348365 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 00:08:30 crc kubenswrapper[4953]: E1203 00:08:30.348597 4953 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 00:08:30 crc kubenswrapper[4953]: I1203 00:08:30.354880 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:30 crc kubenswrapper[4953]: I1203 00:08:30.354941 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:30 crc kubenswrapper[4953]: I1203 00:08:30.354959 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:30 crc kubenswrapper[4953]: I1203 00:08:30.355047 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:30 crc kubenswrapper[4953]: I1203 00:08:30.355071 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:30Z","lastTransitionTime":"2025-12-03T00:08:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:30 crc kubenswrapper[4953]: I1203 00:08:30.458336 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:30 crc kubenswrapper[4953]: I1203 00:08:30.458420 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:30 crc kubenswrapper[4953]: I1203 00:08:30.458443 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:30 crc kubenswrapper[4953]: I1203 00:08:30.458473 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:30 crc kubenswrapper[4953]: I1203 00:08:30.458494 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:30Z","lastTransitionTime":"2025-12-03T00:08:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:30 crc kubenswrapper[4953]: I1203 00:08:30.563163 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:30 crc kubenswrapper[4953]: I1203 00:08:30.563241 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:30 crc kubenswrapper[4953]: I1203 00:08:30.563264 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:30 crc kubenswrapper[4953]: I1203 00:08:30.563296 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:30 crc kubenswrapper[4953]: I1203 00:08:30.563317 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:30Z","lastTransitionTime":"2025-12-03T00:08:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:30 crc kubenswrapper[4953]: I1203 00:08:30.666551 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:30 crc kubenswrapper[4953]: I1203 00:08:30.666633 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:30 crc kubenswrapper[4953]: I1203 00:08:30.666658 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:30 crc kubenswrapper[4953]: I1203 00:08:30.666688 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:30 crc kubenswrapper[4953]: I1203 00:08:30.666711 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:30Z","lastTransitionTime":"2025-12-03T00:08:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:30 crc kubenswrapper[4953]: I1203 00:08:30.770034 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:30 crc kubenswrapper[4953]: I1203 00:08:30.770105 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:30 crc kubenswrapper[4953]: I1203 00:08:30.770126 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:30 crc kubenswrapper[4953]: I1203 00:08:30.770155 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:30 crc kubenswrapper[4953]: I1203 00:08:30.770173 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:30Z","lastTransitionTime":"2025-12-03T00:08:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:30 crc kubenswrapper[4953]: I1203 00:08:30.872701 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:30 crc kubenswrapper[4953]: I1203 00:08:30.872780 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:30 crc kubenswrapper[4953]: I1203 00:08:30.872802 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:30 crc kubenswrapper[4953]: I1203 00:08:30.872832 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:30 crc kubenswrapper[4953]: I1203 00:08:30.872854 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:30Z","lastTransitionTime":"2025-12-03T00:08:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:30 crc kubenswrapper[4953]: I1203 00:08:30.976168 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:30 crc kubenswrapper[4953]: I1203 00:08:30.976234 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:30 crc kubenswrapper[4953]: I1203 00:08:30.976259 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:30 crc kubenswrapper[4953]: I1203 00:08:30.976286 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:30 crc kubenswrapper[4953]: I1203 00:08:30.976307 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:30Z","lastTransitionTime":"2025-12-03T00:08:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:31 crc kubenswrapper[4953]: I1203 00:08:31.078544 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:31 crc kubenswrapper[4953]: I1203 00:08:31.078600 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:31 crc kubenswrapper[4953]: I1203 00:08:31.078610 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:31 crc kubenswrapper[4953]: I1203 00:08:31.078625 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:31 crc kubenswrapper[4953]: I1203 00:08:31.078656 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:31Z","lastTransitionTime":"2025-12-03T00:08:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:31 crc kubenswrapper[4953]: I1203 00:08:31.180911 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:31 crc kubenswrapper[4953]: I1203 00:08:31.180929 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:31 crc kubenswrapper[4953]: I1203 00:08:31.180937 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:31 crc kubenswrapper[4953]: I1203 00:08:31.180948 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:31 crc kubenswrapper[4953]: I1203 00:08:31.180956 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:31Z","lastTransitionTime":"2025-12-03T00:08:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:31 crc kubenswrapper[4953]: I1203 00:08:31.283926 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:31 crc kubenswrapper[4953]: I1203 00:08:31.283960 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:31 crc kubenswrapper[4953]: I1203 00:08:31.283971 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:31 crc kubenswrapper[4953]: I1203 00:08:31.284015 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:31 crc kubenswrapper[4953]: I1203 00:08:31.284033 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:31Z","lastTransitionTime":"2025-12-03T00:08:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:31 crc kubenswrapper[4953]: I1203 00:08:31.347100 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xshk4" Dec 03 00:08:31 crc kubenswrapper[4953]: E1203 00:08:31.347316 4953 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xshk4" podUID="bd03a758-0ca9-47d8-86f7-18b2d610df43" Dec 03 00:08:31 crc kubenswrapper[4953]: I1203 00:08:31.391605 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:31 crc kubenswrapper[4953]: I1203 00:08:31.391667 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:31 crc kubenswrapper[4953]: I1203 00:08:31.391685 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:31 crc kubenswrapper[4953]: I1203 00:08:31.391708 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:31 crc kubenswrapper[4953]: I1203 00:08:31.391725 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:31Z","lastTransitionTime":"2025-12-03T00:08:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:31 crc kubenswrapper[4953]: I1203 00:08:31.447909 4953 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podStartSLOduration=82.447884795 podStartE2EDuration="1m22.447884795s" podCreationTimestamp="2025-12-03 00:07:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 00:08:31.445713321 +0000 UTC m=+100.429994114" watchObservedRunningTime="2025-12-03 00:08:31.447884795 +0000 UTC m=+100.432165608" Dec 03 00:08:31 crc kubenswrapper[4953]: I1203 00:08:31.448266 4953 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-675qk" podStartSLOduration=75.448258644 podStartE2EDuration="1m15.448258644s" podCreationTimestamp="2025-12-03 00:07:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 00:08:31.410718707 +0000 UTC m=+100.394999580" watchObservedRunningTime="2025-12-03 00:08:31.448258644 +0000 UTC m=+100.432539457" Dec 03 00:08:31 crc kubenswrapper[4953]: I1203 00:08:31.467566 4953 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" podStartSLOduration=49.46752449 podStartE2EDuration="49.46752449s" podCreationTimestamp="2025-12-03 00:07:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 00:08:31.467472969 +0000 UTC m=+100.451753772" watchObservedRunningTime="2025-12-03 00:08:31.46752449 +0000 UTC m=+100.451805313" Dec 03 00:08:31 crc kubenswrapper[4953]: I1203 00:08:31.495037 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:31 crc kubenswrapper[4953]: I1203 00:08:31.495080 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:31 crc kubenswrapper[4953]: I1203 00:08:31.495094 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:31 crc kubenswrapper[4953]: I1203 00:08:31.495115 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:31 crc kubenswrapper[4953]: I1203 00:08:31.495131 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:31Z","lastTransitionTime":"2025-12-03T00:08:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:31 crc kubenswrapper[4953]: I1203 00:08:31.521933 4953 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/node-resolver-kl5s9" podStartSLOduration=76.521904754 podStartE2EDuration="1m16.521904754s" podCreationTimestamp="2025-12-03 00:07:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 00:08:31.521225237 +0000 UTC m=+100.505506060" watchObservedRunningTime="2025-12-03 00:08:31.521904754 +0000 UTC m=+100.506185577" Dec 03 00:08:31 crc kubenswrapper[4953]: I1203 00:08:31.522377 4953 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-additional-cni-plugins-wj6rz" podStartSLOduration=75.522363425 podStartE2EDuration="1m15.522363425s" podCreationTimestamp="2025-12-03 00:07:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 00:08:31.510250006 +0000 UTC m=+100.494530809" watchObservedRunningTime="2025-12-03 00:08:31.522363425 +0000 UTC m=+100.506644248" Dec 03 00:08:31 crc kubenswrapper[4953]: I1203 00:08:31.536625 4953 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-chmnz" podStartSLOduration=75.536605597 podStartE2EDuration="1m15.536605597s" podCreationTimestamp="2025-12-03 00:07:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 00:08:31.536604957 +0000 UTC m=+100.520885750" watchObservedRunningTime="2025-12-03 00:08:31.536605597 +0000 UTC m=+100.520886410" Dec 03 00:08:31 crc kubenswrapper[4953]: I1203 00:08:31.567014 4953 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=82.566995748 podStartE2EDuration="1m22.566995748s" podCreationTimestamp="2025-12-03 00:07:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 00:08:31.566893855 +0000 UTC m=+100.551174638" watchObservedRunningTime="2025-12-03 00:08:31.566995748 +0000 UTC m=+100.551276521" Dec 03 00:08:31 crc kubenswrapper[4953]: I1203 00:08:31.598260 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:31 crc kubenswrapper[4953]: I1203 00:08:31.598533 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:31 crc kubenswrapper[4953]: I1203 00:08:31.598631 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:31 crc kubenswrapper[4953]: I1203 00:08:31.598716 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:31 crc kubenswrapper[4953]: I1203 00:08:31.598779 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:31Z","lastTransitionTime":"2025-12-03T00:08:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:31 crc kubenswrapper[4953]: I1203 00:08:31.617682 4953 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-daemon-6dfbn" podStartSLOduration=76.61766656 podStartE2EDuration="1m16.61766656s" podCreationTimestamp="2025-12-03 00:07:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 00:08:31.617227409 +0000 UTC m=+100.601508232" watchObservedRunningTime="2025-12-03 00:08:31.61766656 +0000 UTC m=+100.601947333" Dec 03 00:08:31 crc kubenswrapper[4953]: I1203 00:08:31.661451 4953 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/node-ca-5cmlc" podStartSLOduration=76.661436121 podStartE2EDuration="1m16.661436121s" podCreationTimestamp="2025-12-03 00:07:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 00:08:31.660866937 +0000 UTC m=+100.645147730" watchObservedRunningTime="2025-12-03 00:08:31.661436121 +0000 UTC m=+100.645716904" Dec 03 00:08:31 crc kubenswrapper[4953]: I1203 00:08:31.673610 4953 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" podStartSLOduration=32.673588452 podStartE2EDuration="32.673588452s" podCreationTimestamp="2025-12-03 00:07:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 00:08:31.673022588 +0000 UTC m=+100.657303411" watchObservedRunningTime="2025-12-03 00:08:31.673588452 +0000 UTC m=+100.657869235" Dec 03 00:08:31 crc kubenswrapper[4953]: I1203 00:08:31.700837 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:31 crc kubenswrapper[4953]: I1203 00:08:31.700869 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:31 crc kubenswrapper[4953]: I1203 00:08:31.700879 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:31 crc kubenswrapper[4953]: I1203 00:08:31.700894 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:31 crc kubenswrapper[4953]: I1203 00:08:31.700904 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:31Z","lastTransitionTime":"2025-12-03T00:08:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:31 crc kubenswrapper[4953]: I1203 00:08:31.804069 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:31 crc kubenswrapper[4953]: I1203 00:08:31.804139 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:31 crc kubenswrapper[4953]: I1203 00:08:31.804160 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:31 crc kubenswrapper[4953]: I1203 00:08:31.804189 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:31 crc kubenswrapper[4953]: I1203 00:08:31.804213 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:31Z","lastTransitionTime":"2025-12-03T00:08:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:31 crc kubenswrapper[4953]: I1203 00:08:31.907161 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:31 crc kubenswrapper[4953]: I1203 00:08:31.907220 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:31 crc kubenswrapper[4953]: I1203 00:08:31.907243 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:31 crc kubenswrapper[4953]: I1203 00:08:31.907275 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:31 crc kubenswrapper[4953]: I1203 00:08:31.907296 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:31Z","lastTransitionTime":"2025-12-03T00:08:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:32 crc kubenswrapper[4953]: I1203 00:08:32.016231 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:32 crc kubenswrapper[4953]: I1203 00:08:32.016381 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:32 crc kubenswrapper[4953]: I1203 00:08:32.016481 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:32 crc kubenswrapper[4953]: I1203 00:08:32.016542 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:32 crc kubenswrapper[4953]: I1203 00:08:32.016599 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:32Z","lastTransitionTime":"2025-12-03T00:08:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:32 crc kubenswrapper[4953]: I1203 00:08:32.120220 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:32 crc kubenswrapper[4953]: I1203 00:08:32.120281 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:32 crc kubenswrapper[4953]: I1203 00:08:32.120299 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:32 crc kubenswrapper[4953]: I1203 00:08:32.120323 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:32 crc kubenswrapper[4953]: I1203 00:08:32.120340 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:32Z","lastTransitionTime":"2025-12-03T00:08:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:32 crc kubenswrapper[4953]: I1203 00:08:32.223186 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:32 crc kubenswrapper[4953]: I1203 00:08:32.223266 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:32 crc kubenswrapper[4953]: I1203 00:08:32.223290 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:32 crc kubenswrapper[4953]: I1203 00:08:32.223324 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:32 crc kubenswrapper[4953]: I1203 00:08:32.223347 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:32Z","lastTransitionTime":"2025-12-03T00:08:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:32 crc kubenswrapper[4953]: I1203 00:08:32.325457 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:32 crc kubenswrapper[4953]: I1203 00:08:32.325540 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:32 crc kubenswrapper[4953]: I1203 00:08:32.325572 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:32 crc kubenswrapper[4953]: I1203 00:08:32.325603 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:32 crc kubenswrapper[4953]: I1203 00:08:32.325626 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:32Z","lastTransitionTime":"2025-12-03T00:08:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:32 crc kubenswrapper[4953]: I1203 00:08:32.347098 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 00:08:32 crc kubenswrapper[4953]: I1203 00:08:32.347161 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 00:08:32 crc kubenswrapper[4953]: I1203 00:08:32.347162 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 00:08:32 crc kubenswrapper[4953]: E1203 00:08:32.347284 4953 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 00:08:32 crc kubenswrapper[4953]: E1203 00:08:32.347393 4953 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 00:08:32 crc kubenswrapper[4953]: E1203 00:08:32.347479 4953 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 00:08:32 crc kubenswrapper[4953]: I1203 00:08:32.427868 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:32 crc kubenswrapper[4953]: I1203 00:08:32.427935 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:32 crc kubenswrapper[4953]: I1203 00:08:32.427955 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:32 crc kubenswrapper[4953]: I1203 00:08:32.428011 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:32 crc kubenswrapper[4953]: I1203 00:08:32.428031 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:32Z","lastTransitionTime":"2025-12-03T00:08:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:32 crc kubenswrapper[4953]: I1203 00:08:32.530854 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:32 crc kubenswrapper[4953]: I1203 00:08:32.530928 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:32 crc kubenswrapper[4953]: I1203 00:08:32.530952 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:32 crc kubenswrapper[4953]: I1203 00:08:32.531017 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:32 crc kubenswrapper[4953]: I1203 00:08:32.531045 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:32Z","lastTransitionTime":"2025-12-03T00:08:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:32 crc kubenswrapper[4953]: I1203 00:08:32.634309 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:32 crc kubenswrapper[4953]: I1203 00:08:32.634361 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:32 crc kubenswrapper[4953]: I1203 00:08:32.634373 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:32 crc kubenswrapper[4953]: I1203 00:08:32.634389 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:32 crc kubenswrapper[4953]: I1203 00:08:32.634401 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:32Z","lastTransitionTime":"2025-12-03T00:08:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:32 crc kubenswrapper[4953]: I1203 00:08:32.737763 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:32 crc kubenswrapper[4953]: I1203 00:08:32.737809 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:32 crc kubenswrapper[4953]: I1203 00:08:32.737820 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:32 crc kubenswrapper[4953]: I1203 00:08:32.737839 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:32 crc kubenswrapper[4953]: I1203 00:08:32.737851 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:32Z","lastTransitionTime":"2025-12-03T00:08:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:32 crc kubenswrapper[4953]: I1203 00:08:32.840627 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:32 crc kubenswrapper[4953]: I1203 00:08:32.840696 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:32 crc kubenswrapper[4953]: I1203 00:08:32.840716 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:32 crc kubenswrapper[4953]: I1203 00:08:32.840739 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:32 crc kubenswrapper[4953]: I1203 00:08:32.840757 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:32Z","lastTransitionTime":"2025-12-03T00:08:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:32 crc kubenswrapper[4953]: I1203 00:08:32.943870 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:32 crc kubenswrapper[4953]: I1203 00:08:32.943930 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:32 crc kubenswrapper[4953]: I1203 00:08:32.943952 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:32 crc kubenswrapper[4953]: I1203 00:08:32.944007 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:32 crc kubenswrapper[4953]: I1203 00:08:32.944029 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:32Z","lastTransitionTime":"2025-12-03T00:08:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:33 crc kubenswrapper[4953]: I1203 00:08:33.047334 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:33 crc kubenswrapper[4953]: I1203 00:08:33.047393 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:33 crc kubenswrapper[4953]: I1203 00:08:33.047417 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:33 crc kubenswrapper[4953]: I1203 00:08:33.047447 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:33 crc kubenswrapper[4953]: I1203 00:08:33.047470 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:33Z","lastTransitionTime":"2025-12-03T00:08:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:33 crc kubenswrapper[4953]: I1203 00:08:33.150371 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:33 crc kubenswrapper[4953]: I1203 00:08:33.150432 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:33 crc kubenswrapper[4953]: I1203 00:08:33.150450 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:33 crc kubenswrapper[4953]: I1203 00:08:33.150478 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:33 crc kubenswrapper[4953]: I1203 00:08:33.150495 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:33Z","lastTransitionTime":"2025-12-03T00:08:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:33 crc kubenswrapper[4953]: I1203 00:08:33.253312 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:33 crc kubenswrapper[4953]: I1203 00:08:33.253375 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:33 crc kubenswrapper[4953]: I1203 00:08:33.253394 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:33 crc kubenswrapper[4953]: I1203 00:08:33.253419 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:33 crc kubenswrapper[4953]: I1203 00:08:33.253436 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:33Z","lastTransitionTime":"2025-12-03T00:08:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:33 crc kubenswrapper[4953]: I1203 00:08:33.347504 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xshk4" Dec 03 00:08:33 crc kubenswrapper[4953]: E1203 00:08:33.347718 4953 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xshk4" podUID="bd03a758-0ca9-47d8-86f7-18b2d610df43" Dec 03 00:08:33 crc kubenswrapper[4953]: I1203 00:08:33.355656 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:33 crc kubenswrapper[4953]: I1203 00:08:33.355716 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:33 crc kubenswrapper[4953]: I1203 00:08:33.355739 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:33 crc kubenswrapper[4953]: I1203 00:08:33.355768 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:33 crc kubenswrapper[4953]: I1203 00:08:33.355788 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:33Z","lastTransitionTime":"2025-12-03T00:08:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:33 crc kubenswrapper[4953]: I1203 00:08:33.458498 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:33 crc kubenswrapper[4953]: I1203 00:08:33.458541 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:33 crc kubenswrapper[4953]: I1203 00:08:33.458553 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:33 crc kubenswrapper[4953]: I1203 00:08:33.458572 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:33 crc kubenswrapper[4953]: I1203 00:08:33.458585 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:33Z","lastTransitionTime":"2025-12-03T00:08:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:33 crc kubenswrapper[4953]: I1203 00:08:33.562037 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:33 crc kubenswrapper[4953]: I1203 00:08:33.562311 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:33 crc kubenswrapper[4953]: I1203 00:08:33.562378 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:33 crc kubenswrapper[4953]: I1203 00:08:33.562446 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:33 crc kubenswrapper[4953]: I1203 00:08:33.562528 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:33Z","lastTransitionTime":"2025-12-03T00:08:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:33 crc kubenswrapper[4953]: I1203 00:08:33.666513 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:33 crc kubenswrapper[4953]: I1203 00:08:33.666583 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:33 crc kubenswrapper[4953]: I1203 00:08:33.666603 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:33 crc kubenswrapper[4953]: I1203 00:08:33.666629 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:33 crc kubenswrapper[4953]: I1203 00:08:33.666650 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:33Z","lastTransitionTime":"2025-12-03T00:08:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:33 crc kubenswrapper[4953]: I1203 00:08:33.769702 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:33 crc kubenswrapper[4953]: I1203 00:08:33.769753 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:33 crc kubenswrapper[4953]: I1203 00:08:33.769765 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:33 crc kubenswrapper[4953]: I1203 00:08:33.769789 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:33 crc kubenswrapper[4953]: I1203 00:08:33.769803 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:33Z","lastTransitionTime":"2025-12-03T00:08:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:33 crc kubenswrapper[4953]: I1203 00:08:33.872817 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:33 crc kubenswrapper[4953]: I1203 00:08:33.873289 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:33 crc kubenswrapper[4953]: I1203 00:08:33.873474 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:33 crc kubenswrapper[4953]: I1203 00:08:33.873648 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:33 crc kubenswrapper[4953]: I1203 00:08:33.873833 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:33Z","lastTransitionTime":"2025-12-03T00:08:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:33 crc kubenswrapper[4953]: I1203 00:08:33.977183 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:33 crc kubenswrapper[4953]: I1203 00:08:33.977262 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:33 crc kubenswrapper[4953]: I1203 00:08:33.977279 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:33 crc kubenswrapper[4953]: I1203 00:08:33.977306 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:33 crc kubenswrapper[4953]: I1203 00:08:33.977325 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:33Z","lastTransitionTime":"2025-12-03T00:08:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:34 crc kubenswrapper[4953]: I1203 00:08:34.080720 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:34 crc kubenswrapper[4953]: I1203 00:08:34.080781 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:34 crc kubenswrapper[4953]: I1203 00:08:34.080799 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:34 crc kubenswrapper[4953]: I1203 00:08:34.080824 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:34 crc kubenswrapper[4953]: I1203 00:08:34.080842 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:34Z","lastTransitionTime":"2025-12-03T00:08:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:34 crc kubenswrapper[4953]: I1203 00:08:34.184569 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:34 crc kubenswrapper[4953]: I1203 00:08:34.184630 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:34 crc kubenswrapper[4953]: I1203 00:08:34.184650 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:34 crc kubenswrapper[4953]: I1203 00:08:34.184676 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:34 crc kubenswrapper[4953]: I1203 00:08:34.184692 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:34Z","lastTransitionTime":"2025-12-03T00:08:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:34 crc kubenswrapper[4953]: I1203 00:08:34.287690 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:34 crc kubenswrapper[4953]: I1203 00:08:34.287774 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:34 crc kubenswrapper[4953]: I1203 00:08:34.287799 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:34 crc kubenswrapper[4953]: I1203 00:08:34.287829 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:34 crc kubenswrapper[4953]: I1203 00:08:34.287849 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:34Z","lastTransitionTime":"2025-12-03T00:08:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:34 crc kubenswrapper[4953]: I1203 00:08:34.347637 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 00:08:34 crc kubenswrapper[4953]: I1203 00:08:34.347704 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 00:08:34 crc kubenswrapper[4953]: E1203 00:08:34.347878 4953 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 00:08:34 crc kubenswrapper[4953]: I1203 00:08:34.348025 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 00:08:34 crc kubenswrapper[4953]: E1203 00:08:34.348093 4953 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 00:08:34 crc kubenswrapper[4953]: E1203 00:08:34.348204 4953 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 00:08:34 crc kubenswrapper[4953]: I1203 00:08:34.369373 4953 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd/etcd-crc"] Dec 03 00:08:34 crc kubenswrapper[4953]: I1203 00:08:34.390637 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:34 crc kubenswrapper[4953]: I1203 00:08:34.390723 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:34 crc kubenswrapper[4953]: I1203 00:08:34.390753 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:34 crc kubenswrapper[4953]: I1203 00:08:34.390785 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:34 crc kubenswrapper[4953]: I1203 00:08:34.390807 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:34Z","lastTransitionTime":"2025-12-03T00:08:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:34 crc kubenswrapper[4953]: I1203 00:08:34.495804 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:34 crc kubenswrapper[4953]: I1203 00:08:34.495897 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:34 crc kubenswrapper[4953]: I1203 00:08:34.495924 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:34 crc kubenswrapper[4953]: I1203 00:08:34.495961 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:34 crc kubenswrapper[4953]: I1203 00:08:34.496072 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:34Z","lastTransitionTime":"2025-12-03T00:08:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:34 crc kubenswrapper[4953]: I1203 00:08:34.599133 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:34 crc kubenswrapper[4953]: I1203 00:08:34.599206 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:34 crc kubenswrapper[4953]: I1203 00:08:34.599224 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:34 crc kubenswrapper[4953]: I1203 00:08:34.599249 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:34 crc kubenswrapper[4953]: I1203 00:08:34.599267 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:34Z","lastTransitionTime":"2025-12-03T00:08:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:34 crc kubenswrapper[4953]: I1203 00:08:34.702347 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:34 crc kubenswrapper[4953]: I1203 00:08:34.702441 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:34 crc kubenswrapper[4953]: I1203 00:08:34.702492 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:34 crc kubenswrapper[4953]: I1203 00:08:34.702515 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:34 crc kubenswrapper[4953]: I1203 00:08:34.702534 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:34Z","lastTransitionTime":"2025-12-03T00:08:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:34 crc kubenswrapper[4953]: I1203 00:08:34.805401 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:34 crc kubenswrapper[4953]: I1203 00:08:34.805467 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:34 crc kubenswrapper[4953]: I1203 00:08:34.805485 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:34 crc kubenswrapper[4953]: I1203 00:08:34.805512 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:34 crc kubenswrapper[4953]: I1203 00:08:34.805529 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:34Z","lastTransitionTime":"2025-12-03T00:08:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:34 crc kubenswrapper[4953]: I1203 00:08:34.909544 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:34 crc kubenswrapper[4953]: I1203 00:08:34.909667 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:34 crc kubenswrapper[4953]: I1203 00:08:34.909701 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:34 crc kubenswrapper[4953]: I1203 00:08:34.909735 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:34 crc kubenswrapper[4953]: I1203 00:08:34.909757 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:34Z","lastTransitionTime":"2025-12-03T00:08:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:35 crc kubenswrapper[4953]: I1203 00:08:35.013181 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:35 crc kubenswrapper[4953]: I1203 00:08:35.013236 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:35 crc kubenswrapper[4953]: I1203 00:08:35.013252 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:35 crc kubenswrapper[4953]: I1203 00:08:35.013323 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:35 crc kubenswrapper[4953]: I1203 00:08:35.013351 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:35Z","lastTransitionTime":"2025-12-03T00:08:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:35 crc kubenswrapper[4953]: I1203 00:08:35.102137 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/bd03a758-0ca9-47d8-86f7-18b2d610df43-metrics-certs\") pod \"network-metrics-daemon-xshk4\" (UID: \"bd03a758-0ca9-47d8-86f7-18b2d610df43\") " pod="openshift-multus/network-metrics-daemon-xshk4" Dec 03 00:08:35 crc kubenswrapper[4953]: E1203 00:08:35.102337 4953 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 03 00:08:35 crc kubenswrapper[4953]: E1203 00:08:35.102458 4953 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/bd03a758-0ca9-47d8-86f7-18b2d610df43-metrics-certs podName:bd03a758-0ca9-47d8-86f7-18b2d610df43 nodeName:}" failed. No retries permitted until 2025-12-03 00:09:39.102428825 +0000 UTC m=+168.086709648 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/bd03a758-0ca9-47d8-86f7-18b2d610df43-metrics-certs") pod "network-metrics-daemon-xshk4" (UID: "bd03a758-0ca9-47d8-86f7-18b2d610df43") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 03 00:08:35 crc kubenswrapper[4953]: I1203 00:08:35.116314 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:35 crc kubenswrapper[4953]: I1203 00:08:35.116469 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:35 crc kubenswrapper[4953]: I1203 00:08:35.116496 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:35 crc kubenswrapper[4953]: I1203 00:08:35.116527 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:35 crc kubenswrapper[4953]: I1203 00:08:35.116550 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:35Z","lastTransitionTime":"2025-12-03T00:08:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:35 crc kubenswrapper[4953]: I1203 00:08:35.219716 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:35 crc kubenswrapper[4953]: I1203 00:08:35.219763 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:35 crc kubenswrapper[4953]: I1203 00:08:35.219774 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:35 crc kubenswrapper[4953]: I1203 00:08:35.219792 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:35 crc kubenswrapper[4953]: I1203 00:08:35.219805 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:35Z","lastTransitionTime":"2025-12-03T00:08:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:35 crc kubenswrapper[4953]: I1203 00:08:35.308522 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:08:35 crc kubenswrapper[4953]: I1203 00:08:35.308591 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:08:35 crc kubenswrapper[4953]: I1203 00:08:35.308604 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:08:35 crc kubenswrapper[4953]: I1203 00:08:35.308620 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:08:35 crc kubenswrapper[4953]: I1203 00:08:35.308632 4953 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:08:35Z","lastTransitionTime":"2025-12-03T00:08:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:08:35 crc kubenswrapper[4953]: I1203 00:08:35.348097 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xshk4" Dec 03 00:08:35 crc kubenswrapper[4953]: E1203 00:08:35.348291 4953 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xshk4" podUID="bd03a758-0ca9-47d8-86f7-18b2d610df43" Dec 03 00:08:35 crc kubenswrapper[4953]: I1203 00:08:35.380892 4953 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-version/cluster-version-operator-5c965bbfc6-8rdl7"] Dec 03 00:08:35 crc kubenswrapper[4953]: I1203 00:08:35.381246 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-8rdl7" Dec 03 00:08:35 crc kubenswrapper[4953]: I1203 00:08:35.383643 4953 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Dec 03 00:08:35 crc kubenswrapper[4953]: I1203 00:08:35.384113 4953 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Dec 03 00:08:35 crc kubenswrapper[4953]: I1203 00:08:35.384514 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Dec 03 00:08:35 crc kubenswrapper[4953]: I1203 00:08:35.385022 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Dec 03 00:08:35 crc kubenswrapper[4953]: I1203 00:08:35.426317 4953 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd/etcd-crc" podStartSLOduration=1.4262944659999999 podStartE2EDuration="1.426294466s" podCreationTimestamp="2025-12-03 00:08:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 00:08:35.413301045 +0000 UTC m=+104.397581878" watchObservedRunningTime="2025-12-03 00:08:35.426294466 +0000 UTC m=+104.410575259" Dec 03 00:08:35 crc kubenswrapper[4953]: I1203 00:08:35.506403 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/69bcd8ae-20c5-496f-bf20-2e6580b24262-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-8rdl7\" (UID: \"69bcd8ae-20c5-496f-bf20-2e6580b24262\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-8rdl7" Dec 03 00:08:35 crc kubenswrapper[4953]: I1203 00:08:35.506479 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/69bcd8ae-20c5-496f-bf20-2e6580b24262-service-ca\") pod \"cluster-version-operator-5c965bbfc6-8rdl7\" (UID: \"69bcd8ae-20c5-496f-bf20-2e6580b24262\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-8rdl7" Dec 03 00:08:35 crc kubenswrapper[4953]: I1203 00:08:35.506573 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/69bcd8ae-20c5-496f-bf20-2e6580b24262-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-8rdl7\" (UID: \"69bcd8ae-20c5-496f-bf20-2e6580b24262\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-8rdl7" Dec 03 00:08:35 crc kubenswrapper[4953]: I1203 00:08:35.506639 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/69bcd8ae-20c5-496f-bf20-2e6580b24262-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-8rdl7\" (UID: \"69bcd8ae-20c5-496f-bf20-2e6580b24262\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-8rdl7" Dec 03 00:08:35 crc kubenswrapper[4953]: I1203 00:08:35.506669 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/69bcd8ae-20c5-496f-bf20-2e6580b24262-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-8rdl7\" (UID: \"69bcd8ae-20c5-496f-bf20-2e6580b24262\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-8rdl7" Dec 03 00:08:35 crc kubenswrapper[4953]: I1203 00:08:35.607812 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/69bcd8ae-20c5-496f-bf20-2e6580b24262-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-8rdl7\" (UID: \"69bcd8ae-20c5-496f-bf20-2e6580b24262\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-8rdl7" Dec 03 00:08:35 crc kubenswrapper[4953]: I1203 00:08:35.607884 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/69bcd8ae-20c5-496f-bf20-2e6580b24262-service-ca\") pod \"cluster-version-operator-5c965bbfc6-8rdl7\" (UID: \"69bcd8ae-20c5-496f-bf20-2e6580b24262\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-8rdl7" Dec 03 00:08:35 crc kubenswrapper[4953]: I1203 00:08:35.607966 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/69bcd8ae-20c5-496f-bf20-2e6580b24262-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-8rdl7\" (UID: \"69bcd8ae-20c5-496f-bf20-2e6580b24262\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-8rdl7" Dec 03 00:08:35 crc kubenswrapper[4953]: I1203 00:08:35.608112 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/69bcd8ae-20c5-496f-bf20-2e6580b24262-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-8rdl7\" (UID: \"69bcd8ae-20c5-496f-bf20-2e6580b24262\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-8rdl7" Dec 03 00:08:35 crc kubenswrapper[4953]: I1203 00:08:35.608152 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/69bcd8ae-20c5-496f-bf20-2e6580b24262-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-8rdl7\" (UID: \"69bcd8ae-20c5-496f-bf20-2e6580b24262\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-8rdl7" Dec 03 00:08:35 crc kubenswrapper[4953]: I1203 00:08:35.608273 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/69bcd8ae-20c5-496f-bf20-2e6580b24262-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-8rdl7\" (UID: \"69bcd8ae-20c5-496f-bf20-2e6580b24262\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-8rdl7" Dec 03 00:08:35 crc kubenswrapper[4953]: I1203 00:08:35.608317 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/69bcd8ae-20c5-496f-bf20-2e6580b24262-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-8rdl7\" (UID: \"69bcd8ae-20c5-496f-bf20-2e6580b24262\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-8rdl7" Dec 03 00:08:35 crc kubenswrapper[4953]: I1203 00:08:35.609699 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/69bcd8ae-20c5-496f-bf20-2e6580b24262-service-ca\") pod \"cluster-version-operator-5c965bbfc6-8rdl7\" (UID: \"69bcd8ae-20c5-496f-bf20-2e6580b24262\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-8rdl7" Dec 03 00:08:35 crc kubenswrapper[4953]: I1203 00:08:35.617120 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/69bcd8ae-20c5-496f-bf20-2e6580b24262-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-8rdl7\" (UID: \"69bcd8ae-20c5-496f-bf20-2e6580b24262\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-8rdl7" Dec 03 00:08:35 crc kubenswrapper[4953]: I1203 00:08:35.641381 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/69bcd8ae-20c5-496f-bf20-2e6580b24262-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-8rdl7\" (UID: \"69bcd8ae-20c5-496f-bf20-2e6580b24262\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-8rdl7" Dec 03 00:08:35 crc kubenswrapper[4953]: I1203 00:08:35.695568 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-8rdl7" Dec 03 00:08:35 crc kubenswrapper[4953]: I1203 00:08:35.976532 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-8rdl7" event={"ID":"69bcd8ae-20c5-496f-bf20-2e6580b24262","Type":"ContainerStarted","Data":"c8644c7167a858267732ece6b5ab9dd35eb44767880aed957666516b5bd3fe0b"} Dec 03 00:08:35 crc kubenswrapper[4953]: I1203 00:08:35.976602 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-8rdl7" event={"ID":"69bcd8ae-20c5-496f-bf20-2e6580b24262","Type":"ContainerStarted","Data":"02d7a728e93e56fc6523c5a3321e9014d743f1229348cdaf2516fd1d6824f7ea"} Dec 03 00:08:36 crc kubenswrapper[4953]: I1203 00:08:36.347324 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 00:08:36 crc kubenswrapper[4953]: E1203 00:08:36.347519 4953 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 00:08:36 crc kubenswrapper[4953]: I1203 00:08:36.347524 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 00:08:36 crc kubenswrapper[4953]: I1203 00:08:36.347625 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 00:08:36 crc kubenswrapper[4953]: E1203 00:08:36.347906 4953 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 00:08:36 crc kubenswrapper[4953]: E1203 00:08:36.348047 4953 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 00:08:37 crc kubenswrapper[4953]: I1203 00:08:37.347791 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xshk4" Dec 03 00:08:37 crc kubenswrapper[4953]: E1203 00:08:37.347997 4953 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xshk4" podUID="bd03a758-0ca9-47d8-86f7-18b2d610df43" Dec 03 00:08:38 crc kubenswrapper[4953]: I1203 00:08:38.347810 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 00:08:38 crc kubenswrapper[4953]: I1203 00:08:38.347874 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 00:08:38 crc kubenswrapper[4953]: I1203 00:08:38.348028 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 00:08:38 crc kubenswrapper[4953]: E1203 00:08:38.348045 4953 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 00:08:38 crc kubenswrapper[4953]: E1203 00:08:38.348117 4953 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 00:08:38 crc kubenswrapper[4953]: E1203 00:08:38.348369 4953 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 00:08:38 crc kubenswrapper[4953]: I1203 00:08:38.349152 4953 scope.go:117] "RemoveContainer" containerID="0a4b625eda9b9daff8aa7fe3c5b94e417bcefe08fc6a65b4594f821cfdca7bfd" Dec 03 00:08:38 crc kubenswrapper[4953]: E1203 00:08:38.349347 4953 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-txffx_openshift-ovn-kubernetes(8816ebee-c735-4b8d-b1ec-65349707b347)\"" pod="openshift-ovn-kubernetes/ovnkube-node-txffx" podUID="8816ebee-c735-4b8d-b1ec-65349707b347" Dec 03 00:08:39 crc kubenswrapper[4953]: I1203 00:08:39.347155 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xshk4" Dec 03 00:08:39 crc kubenswrapper[4953]: E1203 00:08:39.347361 4953 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xshk4" podUID="bd03a758-0ca9-47d8-86f7-18b2d610df43" Dec 03 00:08:40 crc kubenswrapper[4953]: I1203 00:08:40.347809 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 00:08:40 crc kubenswrapper[4953]: I1203 00:08:40.347873 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 00:08:40 crc kubenswrapper[4953]: I1203 00:08:40.347825 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 00:08:40 crc kubenswrapper[4953]: E1203 00:08:40.347963 4953 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 00:08:40 crc kubenswrapper[4953]: E1203 00:08:40.348054 4953 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 00:08:40 crc kubenswrapper[4953]: E1203 00:08:40.348123 4953 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 00:08:41 crc kubenswrapper[4953]: I1203 00:08:41.347814 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xshk4" Dec 03 00:08:41 crc kubenswrapper[4953]: E1203 00:08:41.349870 4953 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xshk4" podUID="bd03a758-0ca9-47d8-86f7-18b2d610df43" Dec 03 00:08:42 crc kubenswrapper[4953]: I1203 00:08:42.347333 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 00:08:42 crc kubenswrapper[4953]: I1203 00:08:42.347383 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 00:08:42 crc kubenswrapper[4953]: I1203 00:08:42.347456 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 00:08:42 crc kubenswrapper[4953]: E1203 00:08:42.347494 4953 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 00:08:42 crc kubenswrapper[4953]: E1203 00:08:42.347656 4953 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 00:08:42 crc kubenswrapper[4953]: E1203 00:08:42.347787 4953 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 00:08:43 crc kubenswrapper[4953]: I1203 00:08:43.348418 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xshk4" Dec 03 00:08:43 crc kubenswrapper[4953]: E1203 00:08:43.348854 4953 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xshk4" podUID="bd03a758-0ca9-47d8-86f7-18b2d610df43" Dec 03 00:08:44 crc kubenswrapper[4953]: I1203 00:08:44.347637 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 00:08:44 crc kubenswrapper[4953]: I1203 00:08:44.347670 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 00:08:44 crc kubenswrapper[4953]: I1203 00:08:44.347637 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 00:08:44 crc kubenswrapper[4953]: E1203 00:08:44.347758 4953 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 00:08:44 crc kubenswrapper[4953]: E1203 00:08:44.348129 4953 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 00:08:44 crc kubenswrapper[4953]: E1203 00:08:44.348199 4953 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 00:08:45 crc kubenswrapper[4953]: I1203 00:08:45.347924 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xshk4" Dec 03 00:08:45 crc kubenswrapper[4953]: E1203 00:08:45.348153 4953 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xshk4" podUID="bd03a758-0ca9-47d8-86f7-18b2d610df43" Dec 03 00:08:46 crc kubenswrapper[4953]: I1203 00:08:46.347966 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 00:08:46 crc kubenswrapper[4953]: I1203 00:08:46.348083 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 00:08:46 crc kubenswrapper[4953]: I1203 00:08:46.348083 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 00:08:46 crc kubenswrapper[4953]: E1203 00:08:46.348219 4953 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 00:08:46 crc kubenswrapper[4953]: E1203 00:08:46.348458 4953 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 00:08:46 crc kubenswrapper[4953]: E1203 00:08:46.348644 4953 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 00:08:47 crc kubenswrapper[4953]: I1203 00:08:47.347780 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xshk4" Dec 03 00:08:47 crc kubenswrapper[4953]: E1203 00:08:47.348486 4953 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xshk4" podUID="bd03a758-0ca9-47d8-86f7-18b2d610df43" Dec 03 00:08:48 crc kubenswrapper[4953]: I1203 00:08:48.347195 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 00:08:48 crc kubenswrapper[4953]: I1203 00:08:48.347264 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 00:08:48 crc kubenswrapper[4953]: E1203 00:08:48.347375 4953 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 00:08:48 crc kubenswrapper[4953]: E1203 00:08:48.347500 4953 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 00:08:48 crc kubenswrapper[4953]: I1203 00:08:48.347797 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 00:08:48 crc kubenswrapper[4953]: E1203 00:08:48.348171 4953 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 00:08:49 crc kubenswrapper[4953]: I1203 00:08:49.022915 4953 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-675qk_43a4bad8-71f4-441a-8302-aeb526a986af/kube-multus/1.log" Dec 03 00:08:49 crc kubenswrapper[4953]: I1203 00:08:49.023919 4953 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-675qk_43a4bad8-71f4-441a-8302-aeb526a986af/kube-multus/0.log" Dec 03 00:08:49 crc kubenswrapper[4953]: I1203 00:08:49.023990 4953 generic.go:334] "Generic (PLEG): container finished" podID="43a4bad8-71f4-441a-8302-aeb526a986af" containerID="e701c2ad05014683415ae2fb03b96f3acb38062ba7c5bfc4d0a9e34431695e36" exitCode=1 Dec 03 00:08:49 crc kubenswrapper[4953]: I1203 00:08:49.024036 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-675qk" event={"ID":"43a4bad8-71f4-441a-8302-aeb526a986af","Type":"ContainerDied","Data":"e701c2ad05014683415ae2fb03b96f3acb38062ba7c5bfc4d0a9e34431695e36"} Dec 03 00:08:49 crc kubenswrapper[4953]: I1203 00:08:49.024103 4953 scope.go:117] "RemoveContainer" containerID="5b5794fef6477aab5deed8f642e8cc50b37eb1fd0cc6f19c003f34ad31c1e3c0" Dec 03 00:08:49 crc kubenswrapper[4953]: I1203 00:08:49.024516 4953 scope.go:117] "RemoveContainer" containerID="e701c2ad05014683415ae2fb03b96f3acb38062ba7c5bfc4d0a9e34431695e36" Dec 03 00:08:49 crc kubenswrapper[4953]: E1203 00:08:49.024695 4953 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-multus pod=multus-675qk_openshift-multus(43a4bad8-71f4-441a-8302-aeb526a986af)\"" pod="openshift-multus/multus-675qk" podUID="43a4bad8-71f4-441a-8302-aeb526a986af" Dec 03 00:08:49 crc kubenswrapper[4953]: I1203 00:08:49.043656 4953 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-8rdl7" podStartSLOduration=93.043628051 podStartE2EDuration="1m33.043628051s" podCreationTimestamp="2025-12-03 00:07:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 00:08:35.994492004 +0000 UTC m=+104.978772827" watchObservedRunningTime="2025-12-03 00:08:49.043628051 +0000 UTC m=+118.027908844" Dec 03 00:08:49 crc kubenswrapper[4953]: I1203 00:08:49.347412 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xshk4" Dec 03 00:08:49 crc kubenswrapper[4953]: E1203 00:08:49.347599 4953 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xshk4" podUID="bd03a758-0ca9-47d8-86f7-18b2d610df43" Dec 03 00:08:49 crc kubenswrapper[4953]: I1203 00:08:49.348565 4953 scope.go:117] "RemoveContainer" containerID="0a4b625eda9b9daff8aa7fe3c5b94e417bcefe08fc6a65b4594f821cfdca7bfd" Dec 03 00:08:49 crc kubenswrapper[4953]: E1203 00:08:49.348748 4953 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-txffx_openshift-ovn-kubernetes(8816ebee-c735-4b8d-b1ec-65349707b347)\"" pod="openshift-ovn-kubernetes/ovnkube-node-txffx" podUID="8816ebee-c735-4b8d-b1ec-65349707b347" Dec 03 00:08:50 crc kubenswrapper[4953]: I1203 00:08:50.030154 4953 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-675qk_43a4bad8-71f4-441a-8302-aeb526a986af/kube-multus/1.log" Dec 03 00:08:50 crc kubenswrapper[4953]: I1203 00:08:50.347167 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 00:08:50 crc kubenswrapper[4953]: I1203 00:08:50.347255 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 00:08:50 crc kubenswrapper[4953]: I1203 00:08:50.347167 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 00:08:50 crc kubenswrapper[4953]: E1203 00:08:50.347324 4953 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 00:08:50 crc kubenswrapper[4953]: E1203 00:08:50.347423 4953 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 00:08:50 crc kubenswrapper[4953]: E1203 00:08:50.347518 4953 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 00:08:51 crc kubenswrapper[4953]: E1203 00:08:51.315741 4953 kubelet_node_status.go:497] "Node not becoming ready in time after startup" Dec 03 00:08:51 crc kubenswrapper[4953]: I1203 00:08:51.347233 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xshk4" Dec 03 00:08:51 crc kubenswrapper[4953]: E1203 00:08:51.349261 4953 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xshk4" podUID="bd03a758-0ca9-47d8-86f7-18b2d610df43" Dec 03 00:08:51 crc kubenswrapper[4953]: E1203 00:08:51.430605 4953 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 03 00:08:52 crc kubenswrapper[4953]: I1203 00:08:52.347339 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 00:08:52 crc kubenswrapper[4953]: I1203 00:08:52.347390 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 00:08:52 crc kubenswrapper[4953]: I1203 00:08:52.347370 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 00:08:52 crc kubenswrapper[4953]: E1203 00:08:52.347536 4953 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 00:08:52 crc kubenswrapper[4953]: E1203 00:08:52.347657 4953 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 00:08:52 crc kubenswrapper[4953]: E1203 00:08:52.347847 4953 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 00:08:53 crc kubenswrapper[4953]: I1203 00:08:53.346967 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xshk4" Dec 03 00:08:53 crc kubenswrapper[4953]: E1203 00:08:53.347115 4953 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xshk4" podUID="bd03a758-0ca9-47d8-86f7-18b2d610df43" Dec 03 00:08:54 crc kubenswrapper[4953]: I1203 00:08:54.347803 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 00:08:54 crc kubenswrapper[4953]: I1203 00:08:54.347875 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 00:08:54 crc kubenswrapper[4953]: I1203 00:08:54.348872 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 00:08:54 crc kubenswrapper[4953]: E1203 00:08:54.349399 4953 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 00:08:54 crc kubenswrapper[4953]: E1203 00:08:54.349743 4953 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 00:08:54 crc kubenswrapper[4953]: E1203 00:08:54.349946 4953 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 00:08:55 crc kubenswrapper[4953]: I1203 00:08:55.347893 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xshk4" Dec 03 00:08:55 crc kubenswrapper[4953]: E1203 00:08:55.348167 4953 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xshk4" podUID="bd03a758-0ca9-47d8-86f7-18b2d610df43" Dec 03 00:08:56 crc kubenswrapper[4953]: I1203 00:08:56.347175 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 00:08:56 crc kubenswrapper[4953]: E1203 00:08:56.347390 4953 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 00:08:56 crc kubenswrapper[4953]: I1203 00:08:56.347432 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 00:08:56 crc kubenswrapper[4953]: E1203 00:08:56.347594 4953 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 00:08:56 crc kubenswrapper[4953]: I1203 00:08:56.347866 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 00:08:56 crc kubenswrapper[4953]: E1203 00:08:56.348355 4953 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 00:08:56 crc kubenswrapper[4953]: E1203 00:08:56.431897 4953 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 03 00:08:57 crc kubenswrapper[4953]: I1203 00:08:57.347249 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xshk4" Dec 03 00:08:57 crc kubenswrapper[4953]: E1203 00:08:57.347419 4953 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xshk4" podUID="bd03a758-0ca9-47d8-86f7-18b2d610df43" Dec 03 00:08:58 crc kubenswrapper[4953]: I1203 00:08:58.347591 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 00:08:58 crc kubenswrapper[4953]: I1203 00:08:58.347616 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 00:08:58 crc kubenswrapper[4953]: I1203 00:08:58.347677 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 00:08:58 crc kubenswrapper[4953]: E1203 00:08:58.347743 4953 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 00:08:58 crc kubenswrapper[4953]: E1203 00:08:58.347963 4953 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 00:08:58 crc kubenswrapper[4953]: E1203 00:08:58.348014 4953 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 00:08:59 crc kubenswrapper[4953]: I1203 00:08:59.347905 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xshk4" Dec 03 00:08:59 crc kubenswrapper[4953]: E1203 00:08:59.348144 4953 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xshk4" podUID="bd03a758-0ca9-47d8-86f7-18b2d610df43" Dec 03 00:09:00 crc kubenswrapper[4953]: I1203 00:09:00.347947 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 00:09:00 crc kubenswrapper[4953]: I1203 00:09:00.348026 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 00:09:00 crc kubenswrapper[4953]: I1203 00:09:00.348036 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 00:09:00 crc kubenswrapper[4953]: E1203 00:09:00.349499 4953 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 00:09:00 crc kubenswrapper[4953]: E1203 00:09:00.349591 4953 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 00:09:00 crc kubenswrapper[4953]: I1203 00:09:00.348365 4953 scope.go:117] "RemoveContainer" containerID="e701c2ad05014683415ae2fb03b96f3acb38062ba7c5bfc4d0a9e34431695e36" Dec 03 00:09:00 crc kubenswrapper[4953]: E1203 00:09:00.349943 4953 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 00:09:01 crc kubenswrapper[4953]: I1203 00:09:01.347579 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xshk4" Dec 03 00:09:01 crc kubenswrapper[4953]: E1203 00:09:01.348431 4953 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xshk4" podUID="bd03a758-0ca9-47d8-86f7-18b2d610df43" Dec 03 00:09:01 crc kubenswrapper[4953]: E1203 00:09:01.432861 4953 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 03 00:09:02 crc kubenswrapper[4953]: I1203 00:09:02.075007 4953 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-675qk_43a4bad8-71f4-441a-8302-aeb526a986af/kube-multus/1.log" Dec 03 00:09:02 crc kubenswrapper[4953]: I1203 00:09:02.075071 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-675qk" event={"ID":"43a4bad8-71f4-441a-8302-aeb526a986af","Type":"ContainerStarted","Data":"599a93b23125c53485f61428a83e40ff9255f5e0e3ce3f5f2c0400cbeede0384"} Dec 03 00:09:02 crc kubenswrapper[4953]: I1203 00:09:02.347341 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 00:09:02 crc kubenswrapper[4953]: I1203 00:09:02.347489 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 00:09:02 crc kubenswrapper[4953]: I1203 00:09:02.347382 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 00:09:02 crc kubenswrapper[4953]: E1203 00:09:02.347578 4953 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 00:09:02 crc kubenswrapper[4953]: E1203 00:09:02.347836 4953 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 00:09:02 crc kubenswrapper[4953]: E1203 00:09:02.347904 4953 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 00:09:03 crc kubenswrapper[4953]: I1203 00:09:03.347561 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xshk4" Dec 03 00:09:03 crc kubenswrapper[4953]: E1203 00:09:03.347724 4953 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xshk4" podUID="bd03a758-0ca9-47d8-86f7-18b2d610df43" Dec 03 00:09:03 crc kubenswrapper[4953]: I1203 00:09:03.347932 4953 scope.go:117] "RemoveContainer" containerID="0a4b625eda9b9daff8aa7fe3c5b94e417bcefe08fc6a65b4594f821cfdca7bfd" Dec 03 00:09:04 crc kubenswrapper[4953]: I1203 00:09:04.085331 4953 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-txffx_8816ebee-c735-4b8d-b1ec-65349707b347/ovnkube-controller/3.log" Dec 03 00:09:04 crc kubenswrapper[4953]: I1203 00:09:04.089073 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-txffx" event={"ID":"8816ebee-c735-4b8d-b1ec-65349707b347","Type":"ContainerStarted","Data":"bcde8ca2a20e2afee0f6cb7164e6068aa82f05c0c978ea3cc4c23618fcb57bac"} Dec 03 00:09:04 crc kubenswrapper[4953]: I1203 00:09:04.089634 4953 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-txffx" Dec 03 00:09:04 crc kubenswrapper[4953]: I1203 00:09:04.137179 4953 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-txffx" podStartSLOduration=108.137154584 podStartE2EDuration="1m48.137154584s" podCreationTimestamp="2025-12-03 00:07:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 00:09:04.135834382 +0000 UTC m=+133.120115205" watchObservedRunningTime="2025-12-03 00:09:04.137154584 +0000 UTC m=+133.121435407" Dec 03 00:09:04 crc kubenswrapper[4953]: I1203 00:09:04.182125 4953 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-xshk4"] Dec 03 00:09:04 crc kubenswrapper[4953]: I1203 00:09:04.182263 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xshk4" Dec 03 00:09:04 crc kubenswrapper[4953]: E1203 00:09:04.182408 4953 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xshk4" podUID="bd03a758-0ca9-47d8-86f7-18b2d610df43" Dec 03 00:09:04 crc kubenswrapper[4953]: I1203 00:09:04.347814 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 00:09:04 crc kubenswrapper[4953]: I1203 00:09:04.347835 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 00:09:04 crc kubenswrapper[4953]: I1203 00:09:04.347948 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 00:09:04 crc kubenswrapper[4953]: E1203 00:09:04.348040 4953 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 00:09:04 crc kubenswrapper[4953]: E1203 00:09:04.348293 4953 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 00:09:04 crc kubenswrapper[4953]: E1203 00:09:04.348499 4953 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 00:09:05 crc kubenswrapper[4953]: I1203 00:09:05.347832 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xshk4" Dec 03 00:09:05 crc kubenswrapper[4953]: E1203 00:09:05.348415 4953 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xshk4" podUID="bd03a758-0ca9-47d8-86f7-18b2d610df43" Dec 03 00:09:06 crc kubenswrapper[4953]: I1203 00:09:06.347108 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 00:09:06 crc kubenswrapper[4953]: I1203 00:09:06.347151 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 00:09:06 crc kubenswrapper[4953]: I1203 00:09:06.347109 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 00:09:06 crc kubenswrapper[4953]: E1203 00:09:06.347365 4953 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 00:09:06 crc kubenswrapper[4953]: E1203 00:09:06.347471 4953 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 00:09:06 crc kubenswrapper[4953]: E1203 00:09:06.347614 4953 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 00:09:07 crc kubenswrapper[4953]: I1203 00:09:07.348030 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xshk4" Dec 03 00:09:07 crc kubenswrapper[4953]: I1203 00:09:07.350916 4953 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Dec 03 00:09:07 crc kubenswrapper[4953]: I1203 00:09:07.350932 4953 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Dec 03 00:09:08 crc kubenswrapper[4953]: I1203 00:09:08.347083 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 00:09:08 crc kubenswrapper[4953]: I1203 00:09:08.347148 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 00:09:08 crc kubenswrapper[4953]: I1203 00:09:08.347264 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 00:09:08 crc kubenswrapper[4953]: I1203 00:09:08.350213 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Dec 03 00:09:08 crc kubenswrapper[4953]: I1203 00:09:08.350273 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Dec 03 00:09:08 crc kubenswrapper[4953]: I1203 00:09:08.350277 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Dec 03 00:09:08 crc kubenswrapper[4953]: I1203 00:09:08.350705 4953 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Dec 03 00:09:13 crc kubenswrapper[4953]: I1203 00:09:13.142935 4953 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-txffx" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.104766 4953 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeReady" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.152807 4953 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-2bh27"] Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.153522 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-2bh27" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.156074 4953 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.156156 4953 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-l7sj9"] Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.156594 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-l7sj9" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.158946 4953 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-machine-approver/machine-approver-56656f9798-zmxjk"] Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.159785 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-zmxjk" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.164519 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.164696 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.165153 4953 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.165241 4953 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.165246 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.165602 4953 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console-operator/console-operator-58897d9998-c56d6"] Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.165349 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.165959 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.166065 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.165767 4953 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.165807 4953 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.165866 4953 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.165917 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.166271 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-c56d6" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.166443 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.166420 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.168660 4953 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-45k44"] Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.169262 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-45k44" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.172918 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.175868 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.176186 4953 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.176378 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.179145 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.179845 4953 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-xbwlm"] Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.180776 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-xbwlm" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.181531 4953 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-pruner-29412000-njkvg"] Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.182041 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-pruner-29412000-njkvg" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.184377 4953 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-hzbcl"] Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.185359 4953 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/downloads-7954f5f757-5rzxr"] Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.193190 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-hzbcl" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.200420 4953 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-49rsf"] Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.201584 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-5rzxr" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.209532 4953 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.211593 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-49rsf" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.212830 4953 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.214129 4953 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-vcx7x"] Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.214637 4953 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-52hrr"] Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.215056 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-52hrr" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.214653 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-vcx7x" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.216729 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.218664 4953 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.218739 4953 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.218672 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.221170 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.221340 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.221441 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.221583 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.221677 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.221989 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.222712 4953 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.223028 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.223222 4953 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.223413 4953 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"pruner-dockercfg-p7bcw" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.223605 4953 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.223764 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.223938 4953 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.224304 4953 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.224319 4953 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-t7ljb"] Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.224586 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.224753 4953 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.224946 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.225122 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-t7ljb" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.225217 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.225781 4953 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.225844 4953 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.225911 4953 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-l29cz"] Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.226006 4953 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.226168 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.226284 4953 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.226302 4953 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.226447 4953 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.226604 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.226688 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.226726 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.226731 4953 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-f9d7485db-m2msx"] Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.226766 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.226911 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.226695 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-l29cz" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.227030 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"serviceca" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.227084 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.227137 4953 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.227261 4953 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.227398 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-m2msx" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.227699 4953 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.240268 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.241526 4953 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-t968q"] Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.242284 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-t968q" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.243015 4953 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.244083 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.244107 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.244147 4953 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.244203 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.244242 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.244253 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.244335 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.244354 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.244449 4953 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.244506 4953 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.244732 4953 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.244806 4953 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.244988 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.244763 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.245249 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.245581 4953 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.245829 4953 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.245955 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.246087 4953 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.245833 4953 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.247349 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.247539 4953 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.247908 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.248135 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.248180 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.248283 4953 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.249369 4953 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-l7m9z"] Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.250924 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.251639 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-l7m9z" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.251736 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.251904 4953 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.254325 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.289359 4953 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.289998 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/ca879198-0a45-4505-b845-60d8b0f5bb5d-audit-policies\") pod \"oauth-openshift-558db77b4-45k44\" (UID: \"ca879198-0a45-4505-b845-60d8b0f5bb5d\") " pod="openshift-authentication/oauth-openshift-558db77b4-45k44" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.290023 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/466d51cd-18a2-419a-a749-1ab91c943f13-config\") pod \"apiserver-76f77b778f-hzbcl\" (UID: \"466d51cd-18a2-419a-a749-1ab91c943f13\") " pod="openshift-apiserver/apiserver-76f77b778f-hzbcl" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.290043 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/ca879198-0a45-4505-b845-60d8b0f5bb5d-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-45k44\" (UID: \"ca879198-0a45-4505-b845-60d8b0f5bb5d\") " pod="openshift-authentication/oauth-openshift-558db77b4-45k44" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.290063 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/ca879198-0a45-4505-b845-60d8b0f5bb5d-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-45k44\" (UID: \"ca879198-0a45-4505-b845-60d8b0f5bb5d\") " pod="openshift-authentication/oauth-openshift-558db77b4-45k44" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.290079 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/a68720c3-4f74-4435-91c6-99fd1b561416-serviceca\") pod \"image-pruner-29412000-njkvg\" (UID: \"a68720c3-4f74-4435-91c6-99fd1b561416\") " pod="openshift-image-registry/image-pruner-29412000-njkvg" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.290112 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/ca879198-0a45-4505-b845-60d8b0f5bb5d-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-45k44\" (UID: \"ca879198-0a45-4505-b845-60d8b0f5bb5d\") " pod="openshift-authentication/oauth-openshift-558db77b4-45k44" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.290127 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b33612b4-02d4-4f26-86bb-73a630d1ee58-serving-cert\") pod \"controller-manager-879f6c89f-l7sj9\" (UID: \"b33612b4-02d4-4f26-86bb-73a630d1ee58\") " pod="openshift-controller-manager/controller-manager-879f6c89f-l7sj9" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.290144 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/8c657b07-c109-4475-bdd9-94808f3a58a0-encryption-config\") pod \"apiserver-7bbb656c7d-2bh27\" (UID: \"8c657b07-c109-4475-bdd9-94808f3a58a0\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-2bh27" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.290159 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/8c657b07-c109-4475-bdd9-94808f3a58a0-audit-dir\") pod \"apiserver-7bbb656c7d-2bh27\" (UID: \"8c657b07-c109-4475-bdd9-94808f3a58a0\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-2bh27" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.290173 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/466d51cd-18a2-419a-a749-1ab91c943f13-etcd-client\") pod \"apiserver-76f77b778f-hzbcl\" (UID: \"466d51cd-18a2-419a-a749-1ab91c943f13\") " pod="openshift-apiserver/apiserver-76f77b778f-hzbcl" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.290187 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8c657b07-c109-4475-bdd9-94808f3a58a0-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-2bh27\" (UID: \"8c657b07-c109-4475-bdd9-94808f3a58a0\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-2bh27" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.290202 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/466d51cd-18a2-419a-a749-1ab91c943f13-image-import-ca\") pod \"apiserver-76f77b778f-hzbcl\" (UID: \"466d51cd-18a2-419a-a749-1ab91c943f13\") " pod="openshift-apiserver/apiserver-76f77b778f-hzbcl" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.290216 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s8xxs\" (UniqueName: \"kubernetes.io/projected/0d6c7577-4632-4e4d-88f0-87548a6c640f-kube-api-access-s8xxs\") pod \"cluster-samples-operator-665b6dd947-49rsf\" (UID: \"0d6c7577-4632-4e4d-88f0-87548a6c640f\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-49rsf" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.290233 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ca879198-0a45-4505-b845-60d8b0f5bb5d-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-45k44\" (UID: \"ca879198-0a45-4505-b845-60d8b0f5bb5d\") " pod="openshift-authentication/oauth-openshift-558db77b4-45k44" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.290263 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/496793fd-c6ab-48e4-a2c5-c259593c5d1c-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-xbwlm\" (UID: \"496793fd-c6ab-48e4-a2c5-c259593c5d1c\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-xbwlm" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.290279 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2c2jt\" (UniqueName: \"kubernetes.io/projected/ca879198-0a45-4505-b845-60d8b0f5bb5d-kube-api-access-2c2jt\") pod \"oauth-openshift-558db77b4-45k44\" (UID: \"ca879198-0a45-4505-b845-60d8b0f5bb5d\") " pod="openshift-authentication/oauth-openshift-558db77b4-45k44" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.290294 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x8twd\" (UniqueName: \"kubernetes.io/projected/496793fd-c6ab-48e4-a2c5-c259593c5d1c-kube-api-access-x8twd\") pod \"machine-api-operator-5694c8668f-xbwlm\" (UID: \"496793fd-c6ab-48e4-a2c5-c259593c5d1c\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-xbwlm" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.290308 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/466d51cd-18a2-419a-a749-1ab91c943f13-node-pullsecrets\") pod \"apiserver-76f77b778f-hzbcl\" (UID: \"466d51cd-18a2-419a-a749-1ab91c943f13\") " pod="openshift-apiserver/apiserver-76f77b778f-hzbcl" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.290325 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/466d51cd-18a2-419a-a749-1ab91c943f13-trusted-ca-bundle\") pod \"apiserver-76f77b778f-hzbcl\" (UID: \"466d51cd-18a2-419a-a749-1ab91c943f13\") " pod="openshift-apiserver/apiserver-76f77b778f-hzbcl" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.290342 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/ca879198-0a45-4505-b845-60d8b0f5bb5d-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-45k44\" (UID: \"ca879198-0a45-4505-b845-60d8b0f5bb5d\") " pod="openshift-authentication/oauth-openshift-558db77b4-45k44" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.290360 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/ca879198-0a45-4505-b845-60d8b0f5bb5d-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-45k44\" (UID: \"ca879198-0a45-4505-b845-60d8b0f5bb5d\") " pod="openshift-authentication/oauth-openshift-558db77b4-45k44" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.290376 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/466d51cd-18a2-419a-a749-1ab91c943f13-audit-dir\") pod \"apiserver-76f77b778f-hzbcl\" (UID: \"466d51cd-18a2-419a-a749-1ab91c943f13\") " pod="openshift-apiserver/apiserver-76f77b778f-hzbcl" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.290392 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/b33612b4-02d4-4f26-86bb-73a630d1ee58-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-l7sj9\" (UID: \"b33612b4-02d4-4f26-86bb-73a630d1ee58\") " pod="openshift-controller-manager/controller-manager-879f6c89f-l7sj9" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.290410 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8bd9553a-3081-43b8-a612-cee2a223cf77-config\") pod \"console-operator-58897d9998-c56d6\" (UID: \"8bd9553a-3081-43b8-a612-cee2a223cf77\") " pod="openshift-console-operator/console-operator-58897d9998-c56d6" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.290428 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/496793fd-c6ab-48e4-a2c5-c259593c5d1c-images\") pod \"machine-api-operator-5694c8668f-xbwlm\" (UID: \"496793fd-c6ab-48e4-a2c5-c259593c5d1c\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-xbwlm" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.290444 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2vwkp\" (UniqueName: \"kubernetes.io/projected/b33612b4-02d4-4f26-86bb-73a630d1ee58-kube-api-access-2vwkp\") pod \"controller-manager-879f6c89f-l7sj9\" (UID: \"b33612b4-02d4-4f26-86bb-73a630d1ee58\") " pod="openshift-controller-manager/controller-manager-879f6c89f-l7sj9" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.290468 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/f8f2cf57-42b1-4bb6-8c17-f9e0edb632fa-auth-proxy-config\") pod \"machine-approver-56656f9798-zmxjk\" (UID: \"f8f2cf57-42b1-4bb6-8c17-f9e0edb632fa\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-zmxjk" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.290510 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/ca879198-0a45-4505-b845-60d8b0f5bb5d-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-45k44\" (UID: \"ca879198-0a45-4505-b845-60d8b0f5bb5d\") " pod="openshift-authentication/oauth-openshift-558db77b4-45k44" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.290527 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m8p68\" (UniqueName: \"kubernetes.io/projected/8c657b07-c109-4475-bdd9-94808f3a58a0-kube-api-access-m8p68\") pod \"apiserver-7bbb656c7d-2bh27\" (UID: \"8c657b07-c109-4475-bdd9-94808f3a58a0\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-2bh27" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.290555 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b33612b4-02d4-4f26-86bb-73a630d1ee58-config\") pod \"controller-manager-879f6c89f-l7sj9\" (UID: \"b33612b4-02d4-4f26-86bb-73a630d1ee58\") " pod="openshift-controller-manager/controller-manager-879f6c89f-l7sj9" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.290574 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/ca879198-0a45-4505-b845-60d8b0f5bb5d-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-45k44\" (UID: \"ca879198-0a45-4505-b845-60d8b0f5bb5d\") " pod="openshift-authentication/oauth-openshift-558db77b4-45k44" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.290596 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/ca879198-0a45-4505-b845-60d8b0f5bb5d-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-45k44\" (UID: \"ca879198-0a45-4505-b845-60d8b0f5bb5d\") " pod="openshift-authentication/oauth-openshift-558db77b4-45k44" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.290618 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/ca879198-0a45-4505-b845-60d8b0f5bb5d-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-45k44\" (UID: \"ca879198-0a45-4505-b845-60d8b0f5bb5d\") " pod="openshift-authentication/oauth-openshift-558db77b4-45k44" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.290638 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wtbw4\" (UniqueName: \"kubernetes.io/projected/466d51cd-18a2-419a-a749-1ab91c943f13-kube-api-access-wtbw4\") pod \"apiserver-76f77b778f-hzbcl\" (UID: \"466d51cd-18a2-419a-a749-1ab91c943f13\") " pod="openshift-apiserver/apiserver-76f77b778f-hzbcl" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.290655 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/f8f2cf57-42b1-4bb6-8c17-f9e0edb632fa-machine-approver-tls\") pod \"machine-approver-56656f9798-zmxjk\" (UID: \"f8f2cf57-42b1-4bb6-8c17-f9e0edb632fa\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-zmxjk" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.290670 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8bd9553a-3081-43b8-a612-cee2a223cf77-trusted-ca\") pod \"console-operator-58897d9998-c56d6\" (UID: \"8bd9553a-3081-43b8-a612-cee2a223cf77\") " pod="openshift-console-operator/console-operator-58897d9998-c56d6" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.290683 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/466d51cd-18a2-419a-a749-1ab91c943f13-etcd-serving-ca\") pod \"apiserver-76f77b778f-hzbcl\" (UID: \"466d51cd-18a2-419a-a749-1ab91c943f13\") " pod="openshift-apiserver/apiserver-76f77b778f-hzbcl" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.290698 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496793fd-c6ab-48e4-a2c5-c259593c5d1c-config\") pod \"machine-api-operator-5694c8668f-xbwlm\" (UID: \"496793fd-c6ab-48e4-a2c5-c259593c5d1c\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-xbwlm" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.290715 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t995w\" (UniqueName: \"kubernetes.io/projected/e0e3abc1-91d2-4a6a-aa1c-dc166f30d75a-kube-api-access-t995w\") pod \"downloads-7954f5f757-5rzxr\" (UID: \"e0e3abc1-91d2-4a6a-aa1c-dc166f30d75a\") " pod="openshift-console/downloads-7954f5f757-5rzxr" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.290728 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/8c657b07-c109-4475-bdd9-94808f3a58a0-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-2bh27\" (UID: \"8c657b07-c109-4475-bdd9-94808f3a58a0\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-2bh27" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.290744 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f8f2cf57-42b1-4bb6-8c17-f9e0edb632fa-config\") pod \"machine-approver-56656f9798-zmxjk\" (UID: \"f8f2cf57-42b1-4bb6-8c17-f9e0edb632fa\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-zmxjk" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.290758 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n8s5z\" (UniqueName: \"kubernetes.io/projected/8bd9553a-3081-43b8-a612-cee2a223cf77-kube-api-access-n8s5z\") pod \"console-operator-58897d9998-c56d6\" (UID: \"8bd9553a-3081-43b8-a612-cee2a223cf77\") " pod="openshift-console-operator/console-operator-58897d9998-c56d6" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.290773 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-67dx2\" (UniqueName: \"kubernetes.io/projected/a68720c3-4f74-4435-91c6-99fd1b561416-kube-api-access-67dx2\") pod \"image-pruner-29412000-njkvg\" (UID: \"a68720c3-4f74-4435-91c6-99fd1b561416\") " pod="openshift-image-registry/image-pruner-29412000-njkvg" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.290794 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/8c657b07-c109-4475-bdd9-94808f3a58a0-etcd-client\") pod \"apiserver-7bbb656c7d-2bh27\" (UID: \"8c657b07-c109-4475-bdd9-94808f3a58a0\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-2bh27" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.290814 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/466d51cd-18a2-419a-a749-1ab91c943f13-audit\") pod \"apiserver-76f77b778f-hzbcl\" (UID: \"466d51cd-18a2-419a-a749-1ab91c943f13\") " pod="openshift-apiserver/apiserver-76f77b778f-hzbcl" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.290836 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/ca879198-0a45-4505-b845-60d8b0f5bb5d-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-45k44\" (UID: \"ca879198-0a45-4505-b845-60d8b0f5bb5d\") " pod="openshift-authentication/oauth-openshift-558db77b4-45k44" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.290857 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/466d51cd-18a2-419a-a749-1ab91c943f13-encryption-config\") pod \"apiserver-76f77b778f-hzbcl\" (UID: \"466d51cd-18a2-419a-a749-1ab91c943f13\") " pod="openshift-apiserver/apiserver-76f77b778f-hzbcl" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.290872 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/ca879198-0a45-4505-b845-60d8b0f5bb5d-audit-dir\") pod \"oauth-openshift-558db77b4-45k44\" (UID: \"ca879198-0a45-4505-b845-60d8b0f5bb5d\") " pod="openshift-authentication/oauth-openshift-558db77b4-45k44" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.290887 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b33612b4-02d4-4f26-86bb-73a630d1ee58-client-ca\") pod \"controller-manager-879f6c89f-l7sj9\" (UID: \"b33612b4-02d4-4f26-86bb-73a630d1ee58\") " pod="openshift-controller-manager/controller-manager-879f6c89f-l7sj9" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.290901 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/0d6c7577-4632-4e4d-88f0-87548a6c640f-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-49rsf\" (UID: \"0d6c7577-4632-4e4d-88f0-87548a6c640f\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-49rsf" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.290921 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gb5z2\" (UniqueName: \"kubernetes.io/projected/f8f2cf57-42b1-4bb6-8c17-f9e0edb632fa-kube-api-access-gb5z2\") pod \"machine-approver-56656f9798-zmxjk\" (UID: \"f8f2cf57-42b1-4bb6-8c17-f9e0edb632fa\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-zmxjk" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.290938 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/8c657b07-c109-4475-bdd9-94808f3a58a0-audit-policies\") pod \"apiserver-7bbb656c7d-2bh27\" (UID: \"8c657b07-c109-4475-bdd9-94808f3a58a0\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-2bh27" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.290953 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8bd9553a-3081-43b8-a612-cee2a223cf77-serving-cert\") pod \"console-operator-58897d9998-c56d6\" (UID: \"8bd9553a-3081-43b8-a612-cee2a223cf77\") " pod="openshift-console-operator/console-operator-58897d9998-c56d6" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.291430 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/466d51cd-18a2-419a-a749-1ab91c943f13-serving-cert\") pod \"apiserver-76f77b778f-hzbcl\" (UID: \"466d51cd-18a2-419a-a749-1ab91c943f13\") " pod="openshift-apiserver/apiserver-76f77b778f-hzbcl" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.291452 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8c657b07-c109-4475-bdd9-94808f3a58a0-serving-cert\") pod \"apiserver-7bbb656c7d-2bh27\" (UID: \"8c657b07-c109-4475-bdd9-94808f3a58a0\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-2bh27" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.292426 4953 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-8k8mf"] Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.292947 4953 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-lwq7b"] Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.293329 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-lwq7b" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.294596 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-8k8mf" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.295874 4953 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-lg8jg"] Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.296364 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-lg8jg" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.298518 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.299333 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.299522 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.299672 4953 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.299689 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.302397 4953 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-95vxq"] Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.303028 4953 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.303087 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-95vxq" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.303300 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.303738 4953 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-gsszn"] Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.305164 4953 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-2bh27"] Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.305352 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-gsszn" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.306757 4953 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.308704 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.309027 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.309268 4953 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-gdstj"] Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.310041 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-gdstj" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.311366 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.311417 4953 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-dlqdp"] Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.311965 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-dlqdp" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.313177 4953 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-lhvnr"] Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.314132 4953 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-z5glq"] Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.314556 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-lhvnr" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.314658 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-z5glq" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.317030 4953 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-qv29t"] Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.317694 4953 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-zd9fs"] Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.318156 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-zd9fs" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.318453 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-qv29t" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.319689 4953 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-njtzc"] Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.320368 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-njtzc" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.324167 4953 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-5fdxh"] Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.324858 4953 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-brcxn"] Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.327008 4953 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-l7sj9"] Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.327103 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-brcxn" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.327630 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-5fdxh" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.332561 4953 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-kx9dt"] Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.333647 4953 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-kz9m9"] Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.334187 4953 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-h95rc"] Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.334662 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-h95rc" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.335224 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-kx9dt" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.335308 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-kz9m9" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.339056 4953 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-fcw6n"] Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.340042 4953 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress/router-default-5444994796-r8l2l"] Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.340425 4953 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-vcx7x"] Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.340512 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.340549 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-r8l2l" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.340875 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-fcw6n" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.340894 4953 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.341164 4953 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-xhc8f"] Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.341510 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-xhc8f" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.342411 4953 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-server-gpwzf"] Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.343181 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-gpwzf" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.344538 4953 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412000-mqgxv"] Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.344933 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29412000-mqgxv" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.365381 4953 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-jfp8j"] Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.369266 4953 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.382601 4953 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-hzbcl"] Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.382642 4953 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-45k44"] Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.382732 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-jfp8j" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.384552 4953 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-c56d6"] Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.385302 4953 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/dns-default-qbsjg"] Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.386118 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-qbsjg" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.388206 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.392165 4953 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-gsszn"] Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.392252 4953 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-pruner-29412000-njkvg"] Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.392528 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ca879198-0a45-4505-b845-60d8b0f5bb5d-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-45k44\" (UID: \"ca879198-0a45-4505-b845-60d8b0f5bb5d\") " pod="openshift-authentication/oauth-openshift-558db77b4-45k44" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.392550 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/496793fd-c6ab-48e4-a2c5-c259593c5d1c-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-xbwlm\" (UID: \"496793fd-c6ab-48e4-a2c5-c259593c5d1c\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-xbwlm" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.392573 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mbl6w\" (UniqueName: \"kubernetes.io/projected/6d99040d-450f-4ba8-a96d-16b694ca3636-kube-api-access-mbl6w\") pod \"authentication-operator-69f744f599-vcx7x\" (UID: \"6d99040d-450f-4ba8-a96d-16b694ca3636\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-vcx7x" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.392590 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2c2jt\" (UniqueName: \"kubernetes.io/projected/ca879198-0a45-4505-b845-60d8b0f5bb5d-kube-api-access-2c2jt\") pod \"oauth-openshift-558db77b4-45k44\" (UID: \"ca879198-0a45-4505-b845-60d8b0f5bb5d\") " pod="openshift-authentication/oauth-openshift-558db77b4-45k44" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.392605 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x8twd\" (UniqueName: \"kubernetes.io/projected/496793fd-c6ab-48e4-a2c5-c259593c5d1c-kube-api-access-x8twd\") pod \"machine-api-operator-5694c8668f-xbwlm\" (UID: \"496793fd-c6ab-48e4-a2c5-c259593c5d1c\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-xbwlm" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.392620 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/466d51cd-18a2-419a-a749-1ab91c943f13-node-pullsecrets\") pod \"apiserver-76f77b778f-hzbcl\" (UID: \"466d51cd-18a2-419a-a749-1ab91c943f13\") " pod="openshift-apiserver/apiserver-76f77b778f-hzbcl" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.392634 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/466d51cd-18a2-419a-a749-1ab91c943f13-trusted-ca-bundle\") pod \"apiserver-76f77b778f-hzbcl\" (UID: \"466d51cd-18a2-419a-a749-1ab91c943f13\") " pod="openshift-apiserver/apiserver-76f77b778f-hzbcl" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.392651 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/8a9efbc7-da01-4454-afce-64f004f240fd-images\") pod \"machine-config-operator-74547568cd-lg8jg\" (UID: \"8a9efbc7-da01-4454-afce-64f004f240fd\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-lg8jg" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.392667 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/416a5985-9d2e-43f0-be89-d8be0ae77816-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-gsszn\" (UID: \"416a5985-9d2e-43f0-be89-d8be0ae77816\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-gsszn" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.392681 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/98788d99-bf86-41b8-b71b-0c1edab5dca2-bound-sa-token\") pod \"ingress-operator-5b745b69d9-lhvnr\" (UID: \"98788d99-bf86-41b8-b71b-0c1edab5dca2\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-lhvnr" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.392697 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/ca879198-0a45-4505-b845-60d8b0f5bb5d-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-45k44\" (UID: \"ca879198-0a45-4505-b845-60d8b0f5bb5d\") " pod="openshift-authentication/oauth-openshift-558db77b4-45k44" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.392714 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/ca879198-0a45-4505-b845-60d8b0f5bb5d-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-45k44\" (UID: \"ca879198-0a45-4505-b845-60d8b0f5bb5d\") " pod="openshift-authentication/oauth-openshift-558db77b4-45k44" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.392729 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/466d51cd-18a2-419a-a749-1ab91c943f13-audit-dir\") pod \"apiserver-76f77b778f-hzbcl\" (UID: \"466d51cd-18a2-419a-a749-1ab91c943f13\") " pod="openshift-apiserver/apiserver-76f77b778f-hzbcl" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.392745 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/b33612b4-02d4-4f26-86bb-73a630d1ee58-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-l7sj9\" (UID: \"b33612b4-02d4-4f26-86bb-73a630d1ee58\") " pod="openshift-controller-manager/controller-manager-879f6c89f-l7sj9" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.392760 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8bd9553a-3081-43b8-a612-cee2a223cf77-config\") pod \"console-operator-58897d9998-c56d6\" (UID: \"8bd9553a-3081-43b8-a612-cee2a223cf77\") " pod="openshift-console-operator/console-operator-58897d9998-c56d6" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.392775 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/496793fd-c6ab-48e4-a2c5-c259593c5d1c-images\") pod \"machine-api-operator-5694c8668f-xbwlm\" (UID: \"496793fd-c6ab-48e4-a2c5-c259593c5d1c\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-xbwlm" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.392790 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2vwkp\" (UniqueName: \"kubernetes.io/projected/b33612b4-02d4-4f26-86bb-73a630d1ee58-kube-api-access-2vwkp\") pod \"controller-manager-879f6c89f-l7sj9\" (UID: \"b33612b4-02d4-4f26-86bb-73a630d1ee58\") " pod="openshift-controller-manager/controller-manager-879f6c89f-l7sj9" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.392811 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/f8f2cf57-42b1-4bb6-8c17-f9e0edb632fa-auth-proxy-config\") pod \"machine-approver-56656f9798-zmxjk\" (UID: \"f8f2cf57-42b1-4bb6-8c17-f9e0edb632fa\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-zmxjk" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.392828 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6d99040d-450f-4ba8-a96d-16b694ca3636-service-ca-bundle\") pod \"authentication-operator-69f744f599-vcx7x\" (UID: \"6d99040d-450f-4ba8-a96d-16b694ca3636\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-vcx7x" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.392847 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/ca879198-0a45-4505-b845-60d8b0f5bb5d-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-45k44\" (UID: \"ca879198-0a45-4505-b845-60d8b0f5bb5d\") " pod="openshift-authentication/oauth-openshift-558db77b4-45k44" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.392862 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wr45v\" (UniqueName: \"kubernetes.io/projected/98788d99-bf86-41b8-b71b-0c1edab5dca2-kube-api-access-wr45v\") pod \"ingress-operator-5b745b69d9-lhvnr\" (UID: \"98788d99-bf86-41b8-b71b-0c1edab5dca2\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-lhvnr" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.392880 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m8p68\" (UniqueName: \"kubernetes.io/projected/8c657b07-c109-4475-bdd9-94808f3a58a0-kube-api-access-m8p68\") pod \"apiserver-7bbb656c7d-2bh27\" (UID: \"8c657b07-c109-4475-bdd9-94808f3a58a0\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-2bh27" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.392894 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/8a9efbc7-da01-4454-afce-64f004f240fd-auth-proxy-config\") pod \"machine-config-operator-74547568cd-lg8jg\" (UID: \"8a9efbc7-da01-4454-afce-64f004f240fd\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-lg8jg" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.392909 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9hdq2\" (UniqueName: \"kubernetes.io/projected/8a9efbc7-da01-4454-afce-64f004f240fd-kube-api-access-9hdq2\") pod \"machine-config-operator-74547568cd-lg8jg\" (UID: \"8a9efbc7-da01-4454-afce-64f004f240fd\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-lg8jg" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.392930 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b33612b4-02d4-4f26-86bb-73a630d1ee58-config\") pod \"controller-manager-879f6c89f-l7sj9\" (UID: \"b33612b4-02d4-4f26-86bb-73a630d1ee58\") " pod="openshift-controller-manager/controller-manager-879f6c89f-l7sj9" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.392947 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/ca879198-0a45-4505-b845-60d8b0f5bb5d-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-45k44\" (UID: \"ca879198-0a45-4505-b845-60d8b0f5bb5d\") " pod="openshift-authentication/oauth-openshift-558db77b4-45k44" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.392964 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/ca879198-0a45-4505-b845-60d8b0f5bb5d-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-45k44\" (UID: \"ca879198-0a45-4505-b845-60d8b0f5bb5d\") " pod="openshift-authentication/oauth-openshift-558db77b4-45k44" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.393009 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/ca879198-0a45-4505-b845-60d8b0f5bb5d-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-45k44\" (UID: \"ca879198-0a45-4505-b845-60d8b0f5bb5d\") " pod="openshift-authentication/oauth-openshift-558db77b4-45k44" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.393026 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wtbw4\" (UniqueName: \"kubernetes.io/projected/466d51cd-18a2-419a-a749-1ab91c943f13-kube-api-access-wtbw4\") pod \"apiserver-76f77b778f-hzbcl\" (UID: \"466d51cd-18a2-419a-a749-1ab91c943f13\") " pod="openshift-apiserver/apiserver-76f77b778f-hzbcl" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.393040 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/f8f2cf57-42b1-4bb6-8c17-f9e0edb632fa-machine-approver-tls\") pod \"machine-approver-56656f9798-zmxjk\" (UID: \"f8f2cf57-42b1-4bb6-8c17-f9e0edb632fa\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-zmxjk" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.393055 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8bd9553a-3081-43b8-a612-cee2a223cf77-trusted-ca\") pod \"console-operator-58897d9998-c56d6\" (UID: \"8bd9553a-3081-43b8-a612-cee2a223cf77\") " pod="openshift-console-operator/console-operator-58897d9998-c56d6" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.393070 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/466d51cd-18a2-419a-a749-1ab91c943f13-etcd-serving-ca\") pod \"apiserver-76f77b778f-hzbcl\" (UID: \"466d51cd-18a2-419a-a749-1ab91c943f13\") " pod="openshift-apiserver/apiserver-76f77b778f-hzbcl" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.393085 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496793fd-c6ab-48e4-a2c5-c259593c5d1c-config\") pod \"machine-api-operator-5694c8668f-xbwlm\" (UID: \"496793fd-c6ab-48e4-a2c5-c259593c5d1c\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-xbwlm" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.393101 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6d99040d-450f-4ba8-a96d-16b694ca3636-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-vcx7x\" (UID: \"6d99040d-450f-4ba8-a96d-16b694ca3636\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-vcx7x" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.393117 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t995w\" (UniqueName: \"kubernetes.io/projected/e0e3abc1-91d2-4a6a-aa1c-dc166f30d75a-kube-api-access-t995w\") pod \"downloads-7954f5f757-5rzxr\" (UID: \"e0e3abc1-91d2-4a6a-aa1c-dc166f30d75a\") " pod="openshift-console/downloads-7954f5f757-5rzxr" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.393133 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/8c657b07-c109-4475-bdd9-94808f3a58a0-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-2bh27\" (UID: \"8c657b07-c109-4475-bdd9-94808f3a58a0\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-2bh27" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.393149 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f8f2cf57-42b1-4bb6-8c17-f9e0edb632fa-config\") pod \"machine-approver-56656f9798-zmxjk\" (UID: \"f8f2cf57-42b1-4bb6-8c17-f9e0edb632fa\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-zmxjk" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.393165 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tbvlg\" (UniqueName: \"kubernetes.io/projected/416a5985-9d2e-43f0-be89-d8be0ae77816-kube-api-access-tbvlg\") pod \"multus-admission-controller-857f4d67dd-gsszn\" (UID: \"416a5985-9d2e-43f0-be89-d8be0ae77816\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-gsszn" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.393181 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n8s5z\" (UniqueName: \"kubernetes.io/projected/8bd9553a-3081-43b8-a612-cee2a223cf77-kube-api-access-n8s5z\") pod \"console-operator-58897d9998-c56d6\" (UID: \"8bd9553a-3081-43b8-a612-cee2a223cf77\") " pod="openshift-console-operator/console-operator-58897d9998-c56d6" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.393198 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-67dx2\" (UniqueName: \"kubernetes.io/projected/a68720c3-4f74-4435-91c6-99fd1b561416-kube-api-access-67dx2\") pod \"image-pruner-29412000-njkvg\" (UID: \"a68720c3-4f74-4435-91c6-99fd1b561416\") " pod="openshift-image-registry/image-pruner-29412000-njkvg" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.393214 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/98788d99-bf86-41b8-b71b-0c1edab5dca2-trusted-ca\") pod \"ingress-operator-5b745b69d9-lhvnr\" (UID: \"98788d99-bf86-41b8-b71b-0c1edab5dca2\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-lhvnr" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.393230 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/8c657b07-c109-4475-bdd9-94808f3a58a0-etcd-client\") pod \"apiserver-7bbb656c7d-2bh27\" (UID: \"8c657b07-c109-4475-bdd9-94808f3a58a0\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-2bh27" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.393248 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/466d51cd-18a2-419a-a749-1ab91c943f13-audit\") pod \"apiserver-76f77b778f-hzbcl\" (UID: \"466d51cd-18a2-419a-a749-1ab91c943f13\") " pod="openshift-apiserver/apiserver-76f77b778f-hzbcl" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.393266 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/ca879198-0a45-4505-b845-60d8b0f5bb5d-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-45k44\" (UID: \"ca879198-0a45-4505-b845-60d8b0f5bb5d\") " pod="openshift-authentication/oauth-openshift-558db77b4-45k44" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.393281 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/466d51cd-18a2-419a-a749-1ab91c943f13-encryption-config\") pod \"apiserver-76f77b778f-hzbcl\" (UID: \"466d51cd-18a2-419a-a749-1ab91c943f13\") " pod="openshift-apiserver/apiserver-76f77b778f-hzbcl" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.393296 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/ca879198-0a45-4505-b845-60d8b0f5bb5d-audit-dir\") pod \"oauth-openshift-558db77b4-45k44\" (UID: \"ca879198-0a45-4505-b845-60d8b0f5bb5d\") " pod="openshift-authentication/oauth-openshift-558db77b4-45k44" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.393311 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b33612b4-02d4-4f26-86bb-73a630d1ee58-client-ca\") pod \"controller-manager-879f6c89f-l7sj9\" (UID: \"b33612b4-02d4-4f26-86bb-73a630d1ee58\") " pod="openshift-controller-manager/controller-manager-879f6c89f-l7sj9" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.393327 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/a7477307-0aa0-461d-8fc7-f4f8fc2f7c7b-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-lwq7b\" (UID: \"a7477307-0aa0-461d-8fc7-f4f8fc2f7c7b\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-lwq7b" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.393351 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gb5z2\" (UniqueName: \"kubernetes.io/projected/f8f2cf57-42b1-4bb6-8c17-f9e0edb632fa-kube-api-access-gb5z2\") pod \"machine-approver-56656f9798-zmxjk\" (UID: \"f8f2cf57-42b1-4bb6-8c17-f9e0edb632fa\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-zmxjk" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.393374 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/0d6c7577-4632-4e4d-88f0-87548a6c640f-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-49rsf\" (UID: \"0d6c7577-4632-4e4d-88f0-87548a6c640f\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-49rsf" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.393390 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/8c657b07-c109-4475-bdd9-94808f3a58a0-audit-policies\") pod \"apiserver-7bbb656c7d-2bh27\" (UID: \"8c657b07-c109-4475-bdd9-94808f3a58a0\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-2bh27" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.393407 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8bd9553a-3081-43b8-a612-cee2a223cf77-serving-cert\") pod \"console-operator-58897d9998-c56d6\" (UID: \"8bd9553a-3081-43b8-a612-cee2a223cf77\") " pod="openshift-console-operator/console-operator-58897d9998-c56d6" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.393424 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/466d51cd-18a2-419a-a749-1ab91c943f13-serving-cert\") pod \"apiserver-76f77b778f-hzbcl\" (UID: \"466d51cd-18a2-419a-a749-1ab91c943f13\") " pod="openshift-apiserver/apiserver-76f77b778f-hzbcl" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.393439 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8c657b07-c109-4475-bdd9-94808f3a58a0-serving-cert\") pod \"apiserver-7bbb656c7d-2bh27\" (UID: \"8c657b07-c109-4475-bdd9-94808f3a58a0\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-2bh27" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.393457 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jg6kd\" (UniqueName: \"kubernetes.io/projected/b2f23cca-bfc0-4e7a-bf15-2fc4065e840b-kube-api-access-jg6kd\") pod \"dns-operator-744455d44c-8k8mf\" (UID: \"b2f23cca-bfc0-4e7a-bf15-2fc4065e840b\") " pod="openshift-dns-operator/dns-operator-744455d44c-8k8mf" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.393474 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a7477307-0aa0-461d-8fc7-f4f8fc2f7c7b-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-lwq7b\" (UID: \"a7477307-0aa0-461d-8fc7-f4f8fc2f7c7b\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-lwq7b" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.393494 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/98788d99-bf86-41b8-b71b-0c1edab5dca2-metrics-tls\") pod \"ingress-operator-5b745b69d9-lhvnr\" (UID: \"98788d99-bf86-41b8-b71b-0c1edab5dca2\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-lhvnr" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.393519 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6d99040d-450f-4ba8-a96d-16b694ca3636-config\") pod \"authentication-operator-69f744f599-vcx7x\" (UID: \"6d99040d-450f-4ba8-a96d-16b694ca3636\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-vcx7x" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.393538 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/ca879198-0a45-4505-b845-60d8b0f5bb5d-audit-policies\") pod \"oauth-openshift-558db77b4-45k44\" (UID: \"ca879198-0a45-4505-b845-60d8b0f5bb5d\") " pod="openshift-authentication/oauth-openshift-558db77b4-45k44" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.393555 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/466d51cd-18a2-419a-a749-1ab91c943f13-config\") pod \"apiserver-76f77b778f-hzbcl\" (UID: \"466d51cd-18a2-419a-a749-1ab91c943f13\") " pod="openshift-apiserver/apiserver-76f77b778f-hzbcl" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.393571 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/ca879198-0a45-4505-b845-60d8b0f5bb5d-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-45k44\" (UID: \"ca879198-0a45-4505-b845-60d8b0f5bb5d\") " pod="openshift-authentication/oauth-openshift-558db77b4-45k44" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.393588 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/ca879198-0a45-4505-b845-60d8b0f5bb5d-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-45k44\" (UID: \"ca879198-0a45-4505-b845-60d8b0f5bb5d\") " pod="openshift-authentication/oauth-openshift-558db77b4-45k44" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.393604 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/a68720c3-4f74-4435-91c6-99fd1b561416-serviceca\") pod \"image-pruner-29412000-njkvg\" (UID: \"a68720c3-4f74-4435-91c6-99fd1b561416\") " pod="openshift-image-registry/image-pruner-29412000-njkvg" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.393623 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/b2f23cca-bfc0-4e7a-bf15-2fc4065e840b-metrics-tls\") pod \"dns-operator-744455d44c-8k8mf\" (UID: \"b2f23cca-bfc0-4e7a-bf15-2fc4065e840b\") " pod="openshift-dns-operator/dns-operator-744455d44c-8k8mf" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.393654 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/ca879198-0a45-4505-b845-60d8b0f5bb5d-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-45k44\" (UID: \"ca879198-0a45-4505-b845-60d8b0f5bb5d\") " pod="openshift-authentication/oauth-openshift-558db77b4-45k44" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.393670 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b33612b4-02d4-4f26-86bb-73a630d1ee58-serving-cert\") pod \"controller-manager-879f6c89f-l7sj9\" (UID: \"b33612b4-02d4-4f26-86bb-73a630d1ee58\") " pod="openshift-controller-manager/controller-manager-879f6c89f-l7sj9" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.393684 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/8c657b07-c109-4475-bdd9-94808f3a58a0-encryption-config\") pod \"apiserver-7bbb656c7d-2bh27\" (UID: \"8c657b07-c109-4475-bdd9-94808f3a58a0\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-2bh27" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.393698 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/8c657b07-c109-4475-bdd9-94808f3a58a0-audit-dir\") pod \"apiserver-7bbb656c7d-2bh27\" (UID: \"8c657b07-c109-4475-bdd9-94808f3a58a0\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-2bh27" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.393712 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/8a9efbc7-da01-4454-afce-64f004f240fd-proxy-tls\") pod \"machine-config-operator-74547568cd-lg8jg\" (UID: \"8a9efbc7-da01-4454-afce-64f004f240fd\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-lg8jg" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.393728 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/466d51cd-18a2-419a-a749-1ab91c943f13-etcd-client\") pod \"apiserver-76f77b778f-hzbcl\" (UID: \"466d51cd-18a2-419a-a749-1ab91c943f13\") " pod="openshift-apiserver/apiserver-76f77b778f-hzbcl" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.393743 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a7477307-0aa0-461d-8fc7-f4f8fc2f7c7b-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-lwq7b\" (UID: \"a7477307-0aa0-461d-8fc7-f4f8fc2f7c7b\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-lwq7b" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.393758 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xgl5q\" (UniqueName: \"kubernetes.io/projected/a7477307-0aa0-461d-8fc7-f4f8fc2f7c7b-kube-api-access-xgl5q\") pod \"cluster-image-registry-operator-dc59b4c8b-lwq7b\" (UID: \"a7477307-0aa0-461d-8fc7-f4f8fc2f7c7b\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-lwq7b" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.393774 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/466d51cd-18a2-419a-a749-1ab91c943f13-image-import-ca\") pod \"apiserver-76f77b778f-hzbcl\" (UID: \"466d51cd-18a2-419a-a749-1ab91c943f13\") " pod="openshift-apiserver/apiserver-76f77b778f-hzbcl" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.393791 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s8xxs\" (UniqueName: \"kubernetes.io/projected/0d6c7577-4632-4e4d-88f0-87548a6c640f-kube-api-access-s8xxs\") pod \"cluster-samples-operator-665b6dd947-49rsf\" (UID: \"0d6c7577-4632-4e4d-88f0-87548a6c640f\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-49rsf" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.393808 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8c657b07-c109-4475-bdd9-94808f3a58a0-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-2bh27\" (UID: \"8c657b07-c109-4475-bdd9-94808f3a58a0\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-2bh27" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.393824 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6d99040d-450f-4ba8-a96d-16b694ca3636-serving-cert\") pod \"authentication-operator-69f744f599-vcx7x\" (UID: \"6d99040d-450f-4ba8-a96d-16b694ca3636\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-vcx7x" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.397236 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8bd9553a-3081-43b8-a612-cee2a223cf77-trusted-ca\") pod \"console-operator-58897d9998-c56d6\" (UID: \"8bd9553a-3081-43b8-a612-cee2a223cf77\") " pod="openshift-console-operator/console-operator-58897d9998-c56d6" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.398000 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b33612b4-02d4-4f26-86bb-73a630d1ee58-client-ca\") pod \"controller-manager-879f6c89f-l7sj9\" (UID: \"b33612b4-02d4-4f26-86bb-73a630d1ee58\") " pod="openshift-controller-manager/controller-manager-879f6c89f-l7sj9" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.398424 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/466d51cd-18a2-419a-a749-1ab91c943f13-audit\") pod \"apiserver-76f77b778f-hzbcl\" (UID: \"466d51cd-18a2-419a-a749-1ab91c943f13\") " pod="openshift-apiserver/apiserver-76f77b778f-hzbcl" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.398494 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/466d51cd-18a2-419a-a749-1ab91c943f13-trusted-ca-bundle\") pod \"apiserver-76f77b778f-hzbcl\" (UID: \"466d51cd-18a2-419a-a749-1ab91c943f13\") " pod="openshift-apiserver/apiserver-76f77b778f-hzbcl" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.398518 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8bd9553a-3081-43b8-a612-cee2a223cf77-config\") pod \"console-operator-58897d9998-c56d6\" (UID: \"8bd9553a-3081-43b8-a612-cee2a223cf77\") " pod="openshift-console-operator/console-operator-58897d9998-c56d6" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.399503 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ca879198-0a45-4505-b845-60d8b0f5bb5d-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-45k44\" (UID: \"ca879198-0a45-4505-b845-60d8b0f5bb5d\") " pod="openshift-authentication/oauth-openshift-558db77b4-45k44" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.400256 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/ca879198-0a45-4505-b845-60d8b0f5bb5d-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-45k44\" (UID: \"ca879198-0a45-4505-b845-60d8b0f5bb5d\") " pod="openshift-authentication/oauth-openshift-558db77b4-45k44" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.403546 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b33612b4-02d4-4f26-86bb-73a630d1ee58-serving-cert\") pod \"controller-manager-879f6c89f-l7sj9\" (UID: \"b33612b4-02d4-4f26-86bb-73a630d1ee58\") " pod="openshift-controller-manager/controller-manager-879f6c89f-l7sj9" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.407162 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/8c657b07-c109-4475-bdd9-94808f3a58a0-encryption-config\") pod \"apiserver-7bbb656c7d-2bh27\" (UID: \"8c657b07-c109-4475-bdd9-94808f3a58a0\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-2bh27" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.407218 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/8c657b07-c109-4475-bdd9-94808f3a58a0-audit-dir\") pod \"apiserver-7bbb656c7d-2bh27\" (UID: \"8c657b07-c109-4475-bdd9-94808f3a58a0\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-2bh27" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.408061 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/466d51cd-18a2-419a-a749-1ab91c943f13-etcd-serving-ca\") pod \"apiserver-76f77b778f-hzbcl\" (UID: \"466d51cd-18a2-419a-a749-1ab91c943f13\") " pod="openshift-apiserver/apiserver-76f77b778f-hzbcl" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.408211 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/496793fd-c6ab-48e4-a2c5-c259593c5d1c-images\") pod \"machine-api-operator-5694c8668f-xbwlm\" (UID: \"496793fd-c6ab-48e4-a2c5-c259593c5d1c\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-xbwlm" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.408361 4953 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-5rzxr"] Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.408535 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/466d51cd-18a2-419a-a749-1ab91c943f13-node-pullsecrets\") pod \"apiserver-76f77b778f-hzbcl\" (UID: \"466d51cd-18a2-419a-a749-1ab91c943f13\") " pod="openshift-apiserver/apiserver-76f77b778f-hzbcl" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.408880 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/ca879198-0a45-4505-b845-60d8b0f5bb5d-audit-policies\") pod \"oauth-openshift-558db77b4-45k44\" (UID: \"ca879198-0a45-4505-b845-60d8b0f5bb5d\") " pod="openshift-authentication/oauth-openshift-558db77b4-45k44" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.408922 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f8f2cf57-42b1-4bb6-8c17-f9e0edb632fa-config\") pod \"machine-approver-56656f9798-zmxjk\" (UID: \"f8f2cf57-42b1-4bb6-8c17-f9e0edb632fa\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-zmxjk" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.408931 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/8c657b07-c109-4475-bdd9-94808f3a58a0-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-2bh27\" (UID: \"8c657b07-c109-4475-bdd9-94808f3a58a0\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-2bh27" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.409601 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/f8f2cf57-42b1-4bb6-8c17-f9e0edb632fa-auth-proxy-config\") pod \"machine-approver-56656f9798-zmxjk\" (UID: \"f8f2cf57-42b1-4bb6-8c17-f9e0edb632fa\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-zmxjk" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.411791 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b33612b4-02d4-4f26-86bb-73a630d1ee58-config\") pod \"controller-manager-879f6c89f-l7sj9\" (UID: \"b33612b4-02d4-4f26-86bb-73a630d1ee58\") " pod="openshift-controller-manager/controller-manager-879f6c89f-l7sj9" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.412013 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/466d51cd-18a2-419a-a749-1ab91c943f13-audit-dir\") pod \"apiserver-76f77b778f-hzbcl\" (UID: \"466d51cd-18a2-419a-a749-1ab91c943f13\") " pod="openshift-apiserver/apiserver-76f77b778f-hzbcl" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.412620 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/466d51cd-18a2-419a-a749-1ab91c943f13-image-import-ca\") pod \"apiserver-76f77b778f-hzbcl\" (UID: \"466d51cd-18a2-419a-a749-1ab91c943f13\") " pod="openshift-apiserver/apiserver-76f77b778f-hzbcl" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.414393 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/ca879198-0a45-4505-b845-60d8b0f5bb5d-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-45k44\" (UID: \"ca879198-0a45-4505-b845-60d8b0f5bb5d\") " pod="openshift-authentication/oauth-openshift-558db77b4-45k44" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.415011 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/8c657b07-c109-4475-bdd9-94808f3a58a0-audit-policies\") pod \"apiserver-7bbb656c7d-2bh27\" (UID: \"8c657b07-c109-4475-bdd9-94808f3a58a0\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-2bh27" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.415363 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8c657b07-c109-4475-bdd9-94808f3a58a0-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-2bh27\" (UID: \"8c657b07-c109-4475-bdd9-94808f3a58a0\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-2bh27" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.415394 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/ca879198-0a45-4505-b845-60d8b0f5bb5d-audit-dir\") pod \"oauth-openshift-558db77b4-45k44\" (UID: \"ca879198-0a45-4505-b845-60d8b0f5bb5d\") " pod="openshift-authentication/oauth-openshift-558db77b4-45k44" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.417223 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/466d51cd-18a2-419a-a749-1ab91c943f13-config\") pod \"apiserver-76f77b778f-hzbcl\" (UID: \"466d51cd-18a2-419a-a749-1ab91c943f13\") " pod="openshift-apiserver/apiserver-76f77b778f-hzbcl" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.417626 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/ca879198-0a45-4505-b845-60d8b0f5bb5d-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-45k44\" (UID: \"ca879198-0a45-4505-b845-60d8b0f5bb5d\") " pod="openshift-authentication/oauth-openshift-558db77b4-45k44" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.420856 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/ca879198-0a45-4505-b845-60d8b0f5bb5d-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-45k44\" (UID: \"ca879198-0a45-4505-b845-60d8b0f5bb5d\") " pod="openshift-authentication/oauth-openshift-558db77b4-45k44" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.421761 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/a68720c3-4f74-4435-91c6-99fd1b561416-serviceca\") pod \"image-pruner-29412000-njkvg\" (UID: \"a68720c3-4f74-4435-91c6-99fd1b561416\") " pod="openshift-image-registry/image-pruner-29412000-njkvg" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.424487 4953 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.424707 4953 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.425698 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/b33612b4-02d4-4f26-86bb-73a630d1ee58-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-l7sj9\" (UID: \"b33612b4-02d4-4f26-86bb-73a630d1ee58\") " pod="openshift-controller-manager/controller-manager-879f6c89f-l7sj9" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.426614 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8bd9553a-3081-43b8-a612-cee2a223cf77-serving-cert\") pod \"console-operator-58897d9998-c56d6\" (UID: \"8bd9553a-3081-43b8-a612-cee2a223cf77\") " pod="openshift-console-operator/console-operator-58897d9998-c56d6" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.426674 4953 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-49rsf"] Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.427593 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8c657b07-c109-4475-bdd9-94808f3a58a0-serving-cert\") pod \"apiserver-7bbb656c7d-2bh27\" (UID: \"8c657b07-c109-4475-bdd9-94808f3a58a0\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-2bh27" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.427673 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/466d51cd-18a2-419a-a749-1ab91c943f13-serving-cert\") pod \"apiserver-76f77b778f-hzbcl\" (UID: \"466d51cd-18a2-419a-a749-1ab91c943f13\") " pod="openshift-apiserver/apiserver-76f77b778f-hzbcl" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.427890 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/0d6c7577-4632-4e4d-88f0-87548a6c640f-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-49rsf\" (UID: \"0d6c7577-4632-4e4d-88f0-87548a6c640f\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-49rsf" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.427963 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/8c657b07-c109-4475-bdd9-94808f3a58a0-etcd-client\") pod \"apiserver-7bbb656c7d-2bh27\" (UID: \"8c657b07-c109-4475-bdd9-94808f3a58a0\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-2bh27" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.428014 4953 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-8k8mf"] Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.428341 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/f8f2cf57-42b1-4bb6-8c17-f9e0edb632fa-machine-approver-tls\") pod \"machine-approver-56656f9798-zmxjk\" (UID: \"f8f2cf57-42b1-4bb6-8c17-f9e0edb632fa\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-zmxjk" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.428685 4953 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-l29cz"] Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.429264 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/466d51cd-18a2-419a-a749-1ab91c943f13-encryption-config\") pod \"apiserver-76f77b778f-hzbcl\" (UID: \"466d51cd-18a2-419a-a749-1ab91c943f13\") " pod="openshift-apiserver/apiserver-76f77b778f-hzbcl" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.430843 4953 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-lhvnr"] Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.431233 4953 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-t7ljb"] Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.433517 4953 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-52hrr"] Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.434471 4953 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-95vxq"] Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.435498 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/ca879198-0a45-4505-b845-60d8b0f5bb5d-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-45k44\" (UID: \"ca879198-0a45-4505-b845-60d8b0f5bb5d\") " pod="openshift-authentication/oauth-openshift-558db77b4-45k44" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.435553 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496793fd-c6ab-48e4-a2c5-c259593c5d1c-config\") pod \"machine-api-operator-5694c8668f-xbwlm\" (UID: \"496793fd-c6ab-48e4-a2c5-c259593c5d1c\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-xbwlm" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.435781 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/ca879198-0a45-4505-b845-60d8b0f5bb5d-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-45k44\" (UID: \"ca879198-0a45-4505-b845-60d8b0f5bb5d\") " pod="openshift-authentication/oauth-openshift-558db77b4-45k44" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.436013 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/ca879198-0a45-4505-b845-60d8b0f5bb5d-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-45k44\" (UID: \"ca879198-0a45-4505-b845-60d8b0f5bb5d\") " pod="openshift-authentication/oauth-openshift-558db77b4-45k44" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.436616 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/ca879198-0a45-4505-b845-60d8b0f5bb5d-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-45k44\" (UID: \"ca879198-0a45-4505-b845-60d8b0f5bb5d\") " pod="openshift-authentication/oauth-openshift-558db77b4-45k44" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.436869 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/ca879198-0a45-4505-b845-60d8b0f5bb5d-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-45k44\" (UID: \"ca879198-0a45-4505-b845-60d8b0f5bb5d\") " pod="openshift-authentication/oauth-openshift-558db77b4-45k44" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.436930 4953 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-brcxn"] Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.436939 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/ca879198-0a45-4505-b845-60d8b0f5bb5d-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-45k44\" (UID: \"ca879198-0a45-4505-b845-60d8b0f5bb5d\") " pod="openshift-authentication/oauth-openshift-558db77b4-45k44" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.437474 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/496793fd-c6ab-48e4-a2c5-c259593c5d1c-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-xbwlm\" (UID: \"496793fd-c6ab-48e4-a2c5-c259593c5d1c\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-xbwlm" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.438210 4953 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-njtzc"] Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.439539 4953 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-l7m9z"] Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.440370 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.440525 4953 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-xbwlm"] Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.441478 4953 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-m2msx"] Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.442864 4953 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-lg8jg"] Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.447313 4953 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-t968q"] Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.447345 4953 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-zd9fs"] Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.448760 4953 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-kz9m9"] Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.449917 4953 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-z5glq"] Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.451930 4953 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-dlqdp"] Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.453516 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/466d51cd-18a2-419a-a749-1ab91c943f13-etcd-client\") pod \"apiserver-76f77b778f-hzbcl\" (UID: \"466d51cd-18a2-419a-a749-1ab91c943f13\") " pod="openshift-apiserver/apiserver-76f77b778f-hzbcl" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.453649 4953 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-kx9dt"] Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.454995 4953 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-qv29t"] Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.455131 4953 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.457057 4953 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-5fdxh"] Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.458758 4953 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-lwq7b"] Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.461080 4953 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-fcw6n"] Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.463819 4953 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-gdstj"] Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.466709 4953 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-h95rc"] Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.467784 4953 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-jfp8j"] Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.470146 4953 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-2xt2r"] Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.471134 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-2xt2r" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.474815 4953 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412000-mqgxv"] Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.476086 4953 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-2xt2r"] Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.477192 4953 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-xhc8f"] Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.478417 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.478554 4953 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-qbsjg"] Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.480099 4953 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-canary/ingress-canary-llvsw"] Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.480778 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-llvsw" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.481140 4953 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-llvsw"] Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.494518 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jg6kd\" (UniqueName: \"kubernetes.io/projected/b2f23cca-bfc0-4e7a-bf15-2fc4065e840b-kube-api-access-jg6kd\") pod \"dns-operator-744455d44c-8k8mf\" (UID: \"b2f23cca-bfc0-4e7a-bf15-2fc4065e840b\") " pod="openshift-dns-operator/dns-operator-744455d44c-8k8mf" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.494565 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a7477307-0aa0-461d-8fc7-f4f8fc2f7c7b-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-lwq7b\" (UID: \"a7477307-0aa0-461d-8fc7-f4f8fc2f7c7b\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-lwq7b" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.494601 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/98788d99-bf86-41b8-b71b-0c1edab5dca2-metrics-tls\") pod \"ingress-operator-5b745b69d9-lhvnr\" (UID: \"98788d99-bf86-41b8-b71b-0c1edab5dca2\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-lhvnr" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.494662 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6d99040d-450f-4ba8-a96d-16b694ca3636-config\") pod \"authentication-operator-69f744f599-vcx7x\" (UID: \"6d99040d-450f-4ba8-a96d-16b694ca3636\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-vcx7x" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.494688 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/b2f23cca-bfc0-4e7a-bf15-2fc4065e840b-metrics-tls\") pod \"dns-operator-744455d44c-8k8mf\" (UID: \"b2f23cca-bfc0-4e7a-bf15-2fc4065e840b\") " pod="openshift-dns-operator/dns-operator-744455d44c-8k8mf" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.494710 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/8a9efbc7-da01-4454-afce-64f004f240fd-proxy-tls\") pod \"machine-config-operator-74547568cd-lg8jg\" (UID: \"8a9efbc7-da01-4454-afce-64f004f240fd\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-lg8jg" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.494758 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a7477307-0aa0-461d-8fc7-f4f8fc2f7c7b-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-lwq7b\" (UID: \"a7477307-0aa0-461d-8fc7-f4f8fc2f7c7b\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-lwq7b" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.494783 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xgl5q\" (UniqueName: \"kubernetes.io/projected/a7477307-0aa0-461d-8fc7-f4f8fc2f7c7b-kube-api-access-xgl5q\") pod \"cluster-image-registry-operator-dc59b4c8b-lwq7b\" (UID: \"a7477307-0aa0-461d-8fc7-f4f8fc2f7c7b\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-lwq7b" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.494818 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6d99040d-450f-4ba8-a96d-16b694ca3636-serving-cert\") pod \"authentication-operator-69f744f599-vcx7x\" (UID: \"6d99040d-450f-4ba8-a96d-16b694ca3636\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-vcx7x" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.494841 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mbl6w\" (UniqueName: \"kubernetes.io/projected/6d99040d-450f-4ba8-a96d-16b694ca3636-kube-api-access-mbl6w\") pod \"authentication-operator-69f744f599-vcx7x\" (UID: \"6d99040d-450f-4ba8-a96d-16b694ca3636\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-vcx7x" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.494863 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/8a9efbc7-da01-4454-afce-64f004f240fd-images\") pod \"machine-config-operator-74547568cd-lg8jg\" (UID: \"8a9efbc7-da01-4454-afce-64f004f240fd\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-lg8jg" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.494888 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/416a5985-9d2e-43f0-be89-d8be0ae77816-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-gsszn\" (UID: \"416a5985-9d2e-43f0-be89-d8be0ae77816\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-gsszn" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.494931 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/98788d99-bf86-41b8-b71b-0c1edab5dca2-bound-sa-token\") pod \"ingress-operator-5b745b69d9-lhvnr\" (UID: \"98788d99-bf86-41b8-b71b-0c1edab5dca2\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-lhvnr" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.495076 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6d99040d-450f-4ba8-a96d-16b694ca3636-service-ca-bundle\") pod \"authentication-operator-69f744f599-vcx7x\" (UID: \"6d99040d-450f-4ba8-a96d-16b694ca3636\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-vcx7x" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.495104 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wr45v\" (UniqueName: \"kubernetes.io/projected/98788d99-bf86-41b8-b71b-0c1edab5dca2-kube-api-access-wr45v\") pod \"ingress-operator-5b745b69d9-lhvnr\" (UID: \"98788d99-bf86-41b8-b71b-0c1edab5dca2\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-lhvnr" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.495139 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/8a9efbc7-da01-4454-afce-64f004f240fd-auth-proxy-config\") pod \"machine-config-operator-74547568cd-lg8jg\" (UID: \"8a9efbc7-da01-4454-afce-64f004f240fd\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-lg8jg" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.495169 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9hdq2\" (UniqueName: \"kubernetes.io/projected/8a9efbc7-da01-4454-afce-64f004f240fd-kube-api-access-9hdq2\") pod \"machine-config-operator-74547568cd-lg8jg\" (UID: \"8a9efbc7-da01-4454-afce-64f004f240fd\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-lg8jg" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.495217 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6d99040d-450f-4ba8-a96d-16b694ca3636-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-vcx7x\" (UID: \"6d99040d-450f-4ba8-a96d-16b694ca3636\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-vcx7x" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.495251 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tbvlg\" (UniqueName: \"kubernetes.io/projected/416a5985-9d2e-43f0-be89-d8be0ae77816-kube-api-access-tbvlg\") pod \"multus-admission-controller-857f4d67dd-gsszn\" (UID: \"416a5985-9d2e-43f0-be89-d8be0ae77816\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-gsszn" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.495293 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/98788d99-bf86-41b8-b71b-0c1edab5dca2-trusted-ca\") pod \"ingress-operator-5b745b69d9-lhvnr\" (UID: \"98788d99-bf86-41b8-b71b-0c1edab5dca2\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-lhvnr" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.495417 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/a7477307-0aa0-461d-8fc7-f4f8fc2f7c7b-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-lwq7b\" (UID: \"a7477307-0aa0-461d-8fc7-f4f8fc2f7c7b\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-lwq7b" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.495532 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6d99040d-450f-4ba8-a96d-16b694ca3636-config\") pod \"authentication-operator-69f744f599-vcx7x\" (UID: \"6d99040d-450f-4ba8-a96d-16b694ca3636\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-vcx7x" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.496393 4953 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.496558 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a7477307-0aa0-461d-8fc7-f4f8fc2f7c7b-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-lwq7b\" (UID: \"a7477307-0aa0-461d-8fc7-f4f8fc2f7c7b\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-lwq7b" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.496760 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/8a9efbc7-da01-4454-afce-64f004f240fd-auth-proxy-config\") pod \"machine-config-operator-74547568cd-lg8jg\" (UID: \"8a9efbc7-da01-4454-afce-64f004f240fd\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-lg8jg" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.498186 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6d99040d-450f-4ba8-a96d-16b694ca3636-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-vcx7x\" (UID: \"6d99040d-450f-4ba8-a96d-16b694ca3636\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-vcx7x" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.498554 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6d99040d-450f-4ba8-a96d-16b694ca3636-service-ca-bundle\") pod \"authentication-operator-69f744f599-vcx7x\" (UID: \"6d99040d-450f-4ba8-a96d-16b694ca3636\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-vcx7x" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.499354 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6d99040d-450f-4ba8-a96d-16b694ca3636-serving-cert\") pod \"authentication-operator-69f744f599-vcx7x\" (UID: \"6d99040d-450f-4ba8-a96d-16b694ca3636\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-vcx7x" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.499361 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/b2f23cca-bfc0-4e7a-bf15-2fc4065e840b-metrics-tls\") pod \"dns-operator-744455d44c-8k8mf\" (UID: \"b2f23cca-bfc0-4e7a-bf15-2fc4065e840b\") " pod="openshift-dns-operator/dns-operator-744455d44c-8k8mf" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.501208 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/a7477307-0aa0-461d-8fc7-f4f8fc2f7c7b-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-lwq7b\" (UID: \"a7477307-0aa0-461d-8fc7-f4f8fc2f7c7b\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-lwq7b" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.516015 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.526076 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/8a9efbc7-da01-4454-afce-64f004f240fd-images\") pod \"machine-config-operator-74547568cd-lg8jg\" (UID: \"8a9efbc7-da01-4454-afce-64f004f240fd\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-lg8jg" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.536621 4953 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.548285 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/8a9efbc7-da01-4454-afce-64f004f240fd-proxy-tls\") pod \"machine-config-operator-74547568cd-lg8jg\" (UID: \"8a9efbc7-da01-4454-afce-64f004f240fd\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-lg8jg" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.556162 4953 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.576795 4953 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.595440 4953 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.615699 4953 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.628740 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/416a5985-9d2e-43f0-be89-d8be0ae77816-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-gsszn\" (UID: \"416a5985-9d2e-43f0-be89-d8be0ae77816\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-gsszn" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.655866 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.676395 4953 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.695672 4953 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.715786 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.735760 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.756019 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.776434 4953 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.796555 4953 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.816330 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.835719 4953 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.877078 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.897180 4953 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.916698 4953 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.930328 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/98788d99-bf86-41b8-b71b-0c1edab5dca2-metrics-tls\") pod \"ingress-operator-5b745b69d9-lhvnr\" (UID: \"98788d99-bf86-41b8-b71b-0c1edab5dca2\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-lhvnr" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.942587 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.948246 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/98788d99-bf86-41b8-b71b-0c1edab5dca2-trusted-ca\") pod \"ingress-operator-5b745b69d9-lhvnr\" (UID: \"98788d99-bf86-41b8-b71b-0c1edab5dca2\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-lhvnr" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.956251 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.976561 4953 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Dec 03 00:09:16 crc kubenswrapper[4953]: I1203 00:09:16.996321 4953 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Dec 03 00:09:17 crc kubenswrapper[4953]: I1203 00:09:17.015646 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Dec 03 00:09:17 crc kubenswrapper[4953]: I1203 00:09:17.035654 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Dec 03 00:09:17 crc kubenswrapper[4953]: I1203 00:09:17.055314 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Dec 03 00:09:17 crc kubenswrapper[4953]: I1203 00:09:17.076260 4953 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Dec 03 00:09:17 crc kubenswrapper[4953]: I1203 00:09:17.096206 4953 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Dec 03 00:09:17 crc kubenswrapper[4953]: I1203 00:09:17.116007 4953 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Dec 03 00:09:17 crc kubenswrapper[4953]: I1203 00:09:17.135300 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Dec 03 00:09:17 crc kubenswrapper[4953]: I1203 00:09:17.156844 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Dec 03 00:09:17 crc kubenswrapper[4953]: I1203 00:09:17.176592 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Dec 03 00:09:17 crc kubenswrapper[4953]: I1203 00:09:17.196386 4953 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Dec 03 00:09:17 crc kubenswrapper[4953]: I1203 00:09:17.216645 4953 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Dec 03 00:09:17 crc kubenswrapper[4953]: I1203 00:09:17.235709 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Dec 03 00:09:17 crc kubenswrapper[4953]: I1203 00:09:17.256345 4953 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Dec 03 00:09:17 crc kubenswrapper[4953]: I1203 00:09:17.277006 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Dec 03 00:09:17 crc kubenswrapper[4953]: I1203 00:09:17.297112 4953 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Dec 03 00:09:17 crc kubenswrapper[4953]: I1203 00:09:17.316610 4953 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Dec 03 00:09:17 crc kubenswrapper[4953]: I1203 00:09:17.334401 4953 request.go:700] Waited for 1.001785739s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-storage-version-migrator/configmaps?fieldSelector=metadata.name%3Dkube-root-ca.crt&limit=500&resourceVersion=0 Dec 03 00:09:17 crc kubenswrapper[4953]: I1203 00:09:17.336354 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Dec 03 00:09:17 crc kubenswrapper[4953]: I1203 00:09:17.356300 4953 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Dec 03 00:09:17 crc kubenswrapper[4953]: I1203 00:09:17.376567 4953 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Dec 03 00:09:17 crc kubenswrapper[4953]: I1203 00:09:17.398126 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Dec 03 00:09:17 crc kubenswrapper[4953]: I1203 00:09:17.417311 4953 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Dec 03 00:09:17 crc kubenswrapper[4953]: I1203 00:09:17.436726 4953 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Dec 03 00:09:17 crc kubenswrapper[4953]: I1203 00:09:17.456284 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Dec 03 00:09:17 crc kubenswrapper[4953]: I1203 00:09:17.476581 4953 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Dec 03 00:09:17 crc kubenswrapper[4953]: I1203 00:09:17.504744 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Dec 03 00:09:17 crc kubenswrapper[4953]: I1203 00:09:17.516387 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Dec 03 00:09:17 crc kubenswrapper[4953]: I1203 00:09:17.536215 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Dec 03 00:09:17 crc kubenswrapper[4953]: I1203 00:09:17.557028 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Dec 03 00:09:17 crc kubenswrapper[4953]: I1203 00:09:17.576884 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Dec 03 00:09:17 crc kubenswrapper[4953]: I1203 00:09:17.596459 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Dec 03 00:09:17 crc kubenswrapper[4953]: I1203 00:09:17.617231 4953 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Dec 03 00:09:17 crc kubenswrapper[4953]: I1203 00:09:17.636053 4953 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Dec 03 00:09:17 crc kubenswrapper[4953]: I1203 00:09:17.655648 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Dec 03 00:09:17 crc kubenswrapper[4953]: I1203 00:09:17.676440 4953 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Dec 03 00:09:17 crc kubenswrapper[4953]: I1203 00:09:17.696575 4953 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Dec 03 00:09:17 crc kubenswrapper[4953]: I1203 00:09:17.716113 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Dec 03 00:09:17 crc kubenswrapper[4953]: I1203 00:09:17.735930 4953 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Dec 03 00:09:17 crc kubenswrapper[4953]: I1203 00:09:17.768310 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Dec 03 00:09:17 crc kubenswrapper[4953]: I1203 00:09:17.777524 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Dec 03 00:09:17 crc kubenswrapper[4953]: I1203 00:09:17.796634 4953 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Dec 03 00:09:17 crc kubenswrapper[4953]: I1203 00:09:17.815746 4953 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Dec 03 00:09:17 crc kubenswrapper[4953]: I1203 00:09:17.835169 4953 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Dec 03 00:09:17 crc kubenswrapper[4953]: I1203 00:09:17.856071 4953 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Dec 03 00:09:17 crc kubenswrapper[4953]: I1203 00:09:17.876413 4953 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Dec 03 00:09:17 crc kubenswrapper[4953]: I1203 00:09:17.895833 4953 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Dec 03 00:09:17 crc kubenswrapper[4953]: I1203 00:09:17.916781 4953 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 03 00:09:17 crc kubenswrapper[4953]: I1203 00:09:17.935708 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 03 00:09:17 crc kubenswrapper[4953]: I1203 00:09:17.955847 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Dec 03 00:09:17 crc kubenswrapper[4953]: I1203 00:09:17.977141 4953 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Dec 03 00:09:17 crc kubenswrapper[4953]: I1203 00:09:17.995298 4953 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Dec 03 00:09:18 crc kubenswrapper[4953]: I1203 00:09:18.016840 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Dec 03 00:09:18 crc kubenswrapper[4953]: I1203 00:09:18.035895 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Dec 03 00:09:18 crc kubenswrapper[4953]: I1203 00:09:18.057575 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Dec 03 00:09:18 crc kubenswrapper[4953]: I1203 00:09:18.076797 4953 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Dec 03 00:09:18 crc kubenswrapper[4953]: I1203 00:09:18.096404 4953 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Dec 03 00:09:18 crc kubenswrapper[4953]: I1203 00:09:18.146428 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t995w\" (UniqueName: \"kubernetes.io/projected/e0e3abc1-91d2-4a6a-aa1c-dc166f30d75a-kube-api-access-t995w\") pod \"downloads-7954f5f757-5rzxr\" (UID: \"e0e3abc1-91d2-4a6a-aa1c-dc166f30d75a\") " pod="openshift-console/downloads-7954f5f757-5rzxr" Dec 03 00:09:18 crc kubenswrapper[4953]: I1203 00:09:18.160205 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n8s5z\" (UniqueName: \"kubernetes.io/projected/8bd9553a-3081-43b8-a612-cee2a223cf77-kube-api-access-n8s5z\") pod \"console-operator-58897d9998-c56d6\" (UID: \"8bd9553a-3081-43b8-a612-cee2a223cf77\") " pod="openshift-console-operator/console-operator-58897d9998-c56d6" Dec 03 00:09:18 crc kubenswrapper[4953]: I1203 00:09:18.179190 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2vwkp\" (UniqueName: \"kubernetes.io/projected/b33612b4-02d4-4f26-86bb-73a630d1ee58-kube-api-access-2vwkp\") pod \"controller-manager-879f6c89f-l7sj9\" (UID: \"b33612b4-02d4-4f26-86bb-73a630d1ee58\") " pod="openshift-controller-manager/controller-manager-879f6c89f-l7sj9" Dec 03 00:09:18 crc kubenswrapper[4953]: I1203 00:09:18.196178 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-67dx2\" (UniqueName: \"kubernetes.io/projected/a68720c3-4f74-4435-91c6-99fd1b561416-kube-api-access-67dx2\") pod \"image-pruner-29412000-njkvg\" (UID: \"a68720c3-4f74-4435-91c6-99fd1b561416\") " pod="openshift-image-registry/image-pruner-29412000-njkvg" Dec 03 00:09:18 crc kubenswrapper[4953]: I1203 00:09:18.196455 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-5rzxr" Dec 03 00:09:18 crc kubenswrapper[4953]: I1203 00:09:18.222704 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wtbw4\" (UniqueName: \"kubernetes.io/projected/466d51cd-18a2-419a-a749-1ab91c943f13-kube-api-access-wtbw4\") pod \"apiserver-76f77b778f-hzbcl\" (UID: \"466d51cd-18a2-419a-a749-1ab91c943f13\") " pod="openshift-apiserver/apiserver-76f77b778f-hzbcl" Dec 03 00:09:18 crc kubenswrapper[4953]: I1203 00:09:18.236860 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m8p68\" (UniqueName: \"kubernetes.io/projected/8c657b07-c109-4475-bdd9-94808f3a58a0-kube-api-access-m8p68\") pod \"apiserver-7bbb656c7d-2bh27\" (UID: \"8c657b07-c109-4475-bdd9-94808f3a58a0\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-2bh27" Dec 03 00:09:18 crc kubenswrapper[4953]: I1203 00:09:18.268177 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s8xxs\" (UniqueName: \"kubernetes.io/projected/0d6c7577-4632-4e4d-88f0-87548a6c640f-kube-api-access-s8xxs\") pod \"cluster-samples-operator-665b6dd947-49rsf\" (UID: \"0d6c7577-4632-4e4d-88f0-87548a6c640f\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-49rsf" Dec 03 00:09:18 crc kubenswrapper[4953]: I1203 00:09:18.277736 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gb5z2\" (UniqueName: \"kubernetes.io/projected/f8f2cf57-42b1-4bb6-8c17-f9e0edb632fa-kube-api-access-gb5z2\") pod \"machine-approver-56656f9798-zmxjk\" (UID: \"f8f2cf57-42b1-4bb6-8c17-f9e0edb632fa\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-zmxjk" Dec 03 00:09:18 crc kubenswrapper[4953]: I1203 00:09:18.299293 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2c2jt\" (UniqueName: \"kubernetes.io/projected/ca879198-0a45-4505-b845-60d8b0f5bb5d-kube-api-access-2c2jt\") pod \"oauth-openshift-558db77b4-45k44\" (UID: \"ca879198-0a45-4505-b845-60d8b0f5bb5d\") " pod="openshift-authentication/oauth-openshift-558db77b4-45k44" Dec 03 00:09:18 crc kubenswrapper[4953]: I1203 00:09:18.311667 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-2bh27" Dec 03 00:09:18 crc kubenswrapper[4953]: I1203 00:09:18.315858 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Dec 03 00:09:18 crc kubenswrapper[4953]: I1203 00:09:18.315878 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x8twd\" (UniqueName: \"kubernetes.io/projected/496793fd-c6ab-48e4-a2c5-c259593c5d1c-kube-api-access-x8twd\") pod \"machine-api-operator-5694c8668f-xbwlm\" (UID: \"496793fd-c6ab-48e4-a2c5-c259593c5d1c\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-xbwlm" Dec 03 00:09:18 crc kubenswrapper[4953]: I1203 00:09:18.319450 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 00:09:18 crc kubenswrapper[4953]: E1203 00:09:18.319551 4953 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 00:11:20.319527553 +0000 UTC m=+269.303808346 (durationBeforeRetry 2m2s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 00:09:18 crc kubenswrapper[4953]: I1203 00:09:18.319999 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 00:09:18 crc kubenswrapper[4953]: I1203 00:09:18.320143 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 00:09:18 crc kubenswrapper[4953]: I1203 00:09:18.320311 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 00:09:18 crc kubenswrapper[4953]: I1203 00:09:18.320752 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 00:09:18 crc kubenswrapper[4953]: I1203 00:09:18.321076 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 00:09:18 crc kubenswrapper[4953]: I1203 00:09:18.325470 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 00:09:18 crc kubenswrapper[4953]: I1203 00:09:18.325815 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 00:09:18 crc kubenswrapper[4953]: I1203 00:09:18.326101 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 00:09:18 crc kubenswrapper[4953]: I1203 00:09:18.338255 4953 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Dec 03 00:09:18 crc kubenswrapper[4953]: I1203 00:09:18.342352 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-l7sj9" Dec 03 00:09:18 crc kubenswrapper[4953]: I1203 00:09:18.351366 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-zmxjk" Dec 03 00:09:18 crc kubenswrapper[4953]: I1203 00:09:18.359102 4953 request.go:700] Waited for 1.887686599s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/hostpath-provisioner/configmaps?fieldSelector=metadata.name%3Dkube-root-ca.crt&limit=500&resourceVersion=0 Dec 03 00:09:18 crc kubenswrapper[4953]: I1203 00:09:18.360938 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Dec 03 00:09:18 crc kubenswrapper[4953]: I1203 00:09:18.383134 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Dec 03 00:09:18 crc kubenswrapper[4953]: I1203 00:09:18.396946 4953 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Dec 03 00:09:18 crc kubenswrapper[4953]: I1203 00:09:18.403590 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-c56d6" Dec 03 00:09:18 crc kubenswrapper[4953]: I1203 00:09:18.405086 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-45k44" Dec 03 00:09:18 crc kubenswrapper[4953]: I1203 00:09:18.413901 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-xbwlm" Dec 03 00:09:18 crc kubenswrapper[4953]: I1203 00:09:18.415678 4953 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Dec 03 00:09:18 crc kubenswrapper[4953]: I1203 00:09:18.447689 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Dec 03 00:09:18 crc kubenswrapper[4953]: I1203 00:09:18.455654 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-pruner-29412000-njkvg" Dec 03 00:09:18 crc kubenswrapper[4953]: I1203 00:09:18.489622 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-hzbcl" Dec 03 00:09:18 crc kubenswrapper[4953]: I1203 00:09:18.491392 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a7477307-0aa0-461d-8fc7-f4f8fc2f7c7b-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-lwq7b\" (UID: \"a7477307-0aa0-461d-8fc7-f4f8fc2f7c7b\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-lwq7b" Dec 03 00:09:18 crc kubenswrapper[4953]: I1203 00:09:18.498763 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jg6kd\" (UniqueName: \"kubernetes.io/projected/b2f23cca-bfc0-4e7a-bf15-2fc4065e840b-kube-api-access-jg6kd\") pod \"dns-operator-744455d44c-8k8mf\" (UID: \"b2f23cca-bfc0-4e7a-bf15-2fc4065e840b\") " pod="openshift-dns-operator/dns-operator-744455d44c-8k8mf" Dec 03 00:09:18 crc kubenswrapper[4953]: I1203 00:09:18.515731 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xgl5q\" (UniqueName: \"kubernetes.io/projected/a7477307-0aa0-461d-8fc7-f4f8fc2f7c7b-kube-api-access-xgl5q\") pod \"cluster-image-registry-operator-dc59b4c8b-lwq7b\" (UID: \"a7477307-0aa0-461d-8fc7-f4f8fc2f7c7b\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-lwq7b" Dec 03 00:09:18 crc kubenswrapper[4953]: I1203 00:09:18.531299 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mbl6w\" (UniqueName: \"kubernetes.io/projected/6d99040d-450f-4ba8-a96d-16b694ca3636-kube-api-access-mbl6w\") pod \"authentication-operator-69f744f599-vcx7x\" (UID: \"6d99040d-450f-4ba8-a96d-16b694ca3636\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-vcx7x" Dec 03 00:09:18 crc kubenswrapper[4953]: I1203 00:09:18.552408 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wr45v\" (UniqueName: \"kubernetes.io/projected/98788d99-bf86-41b8-b71b-0c1edab5dca2-kube-api-access-wr45v\") pod \"ingress-operator-5b745b69d9-lhvnr\" (UID: \"98788d99-bf86-41b8-b71b-0c1edab5dca2\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-lhvnr" Dec 03 00:09:18 crc kubenswrapper[4953]: I1203 00:09:18.559829 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-49rsf" Dec 03 00:09:18 crc kubenswrapper[4953]: I1203 00:09:18.572632 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tbvlg\" (UniqueName: \"kubernetes.io/projected/416a5985-9d2e-43f0-be89-d8be0ae77816-kube-api-access-tbvlg\") pod \"multus-admission-controller-857f4d67dd-gsszn\" (UID: \"416a5985-9d2e-43f0-be89-d8be0ae77816\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-gsszn" Dec 03 00:09:18 crc kubenswrapper[4953]: I1203 00:09:18.573371 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 00:09:18 crc kubenswrapper[4953]: I1203 00:09:18.582901 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-vcx7x" Dec 03 00:09:18 crc kubenswrapper[4953]: I1203 00:09:18.590784 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 00:09:18 crc kubenswrapper[4953]: I1203 00:09:18.591082 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/98788d99-bf86-41b8-b71b-0c1edab5dca2-bound-sa-token\") pod \"ingress-operator-5b745b69d9-lhvnr\" (UID: \"98788d99-bf86-41b8-b71b-0c1edab5dca2\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-lhvnr" Dec 03 00:09:18 crc kubenswrapper[4953]: I1203 00:09:18.608208 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 00:09:18 crc kubenswrapper[4953]: I1203 00:09:18.610218 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9hdq2\" (UniqueName: \"kubernetes.io/projected/8a9efbc7-da01-4454-afce-64f004f240fd-kube-api-access-9hdq2\") pod \"machine-config-operator-74547568cd-lg8jg\" (UID: \"8a9efbc7-da01-4454-afce-64f004f240fd\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-lg8jg" Dec 03 00:09:18 crc kubenswrapper[4953]: I1203 00:09:18.629937 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-lwq7b" Dec 03 00:09:18 crc kubenswrapper[4953]: I1203 00:09:18.635155 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-8k8mf" Dec 03 00:09:18 crc kubenswrapper[4953]: I1203 00:09:18.648325 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-lg8jg" Dec 03 00:09:18 crc kubenswrapper[4953]: I1203 00:09:18.654082 4953 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-l7sj9"] Dec 03 00:09:18 crc kubenswrapper[4953]: I1203 00:09:18.675675 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-lhvnr" Dec 03 00:09:18 crc kubenswrapper[4953]: I1203 00:09:18.675725 4953 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-5rzxr"] Dec 03 00:09:18 crc kubenswrapper[4953]: W1203 00:09:18.714581 4953 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb33612b4_02d4_4f26_86bb_73a630d1ee58.slice/crio-5a263737003b51158ff1c32ac37bbf8e444695110c8cf6d655006e58fc7c0938 WatchSource:0}: Error finding container 5a263737003b51158ff1c32ac37bbf8e444695110c8cf6d655006e58fc7c0938: Status 404 returned error can't find the container with id 5a263737003b51158ff1c32ac37bbf8e444695110c8cf6d655006e58fc7c0938 Dec 03 00:09:18 crc kubenswrapper[4953]: W1203 00:09:18.720417 4953 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode0e3abc1_91d2_4a6a_aa1c_dc166f30d75a.slice/crio-f77fc37ba4f9a607b3950e2f7018608a9c1cbce7dd845f6550acb1498eac8f2a WatchSource:0}: Error finding container f77fc37ba4f9a607b3950e2f7018608a9c1cbce7dd845f6550acb1498eac8f2a: Status 404 returned error can't find the container with id f77fc37ba4f9a607b3950e2f7018608a9c1cbce7dd845f6550acb1498eac8f2a Dec 03 00:09:18 crc kubenswrapper[4953]: I1203 00:09:18.726365 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/4313395c-8501-4b5b-a4cb-2fc986fced03-bound-sa-token\") pod \"image-registry-697d97f7c8-t7ljb\" (UID: \"4313395c-8501-4b5b-a4cb-2fc986fced03\") " pod="openshift-image-registry/image-registry-697d97f7c8-t7ljb" Dec 03 00:09:18 crc kubenswrapper[4953]: I1203 00:09:18.726411 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/4ce6dbe8-0292-4172-81a3-f5cdfdba6120-console-serving-cert\") pod \"console-f9d7485db-m2msx\" (UID: \"4ce6dbe8-0292-4172-81a3-f5cdfdba6120\") " pod="openshift-console/console-f9d7485db-m2msx" Dec 03 00:09:18 crc kubenswrapper[4953]: I1203 00:09:18.726451 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/4ce6dbe8-0292-4172-81a3-f5cdfdba6120-console-oauth-config\") pod \"console-f9d7485db-m2msx\" (UID: \"4ce6dbe8-0292-4172-81a3-f5cdfdba6120\") " pod="openshift-console/console-f9d7485db-m2msx" Dec 03 00:09:18 crc kubenswrapper[4953]: I1203 00:09:18.726500 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/4ce6dbe8-0292-4172-81a3-f5cdfdba6120-service-ca\") pod \"console-f9d7485db-m2msx\" (UID: \"4ce6dbe8-0292-4172-81a3-f5cdfdba6120\") " pod="openshift-console/console-f9d7485db-m2msx" Dec 03 00:09:18 crc kubenswrapper[4953]: I1203 00:09:18.726523 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n5wnd\" (UniqueName: \"kubernetes.io/projected/4ce6dbe8-0292-4172-81a3-f5cdfdba6120-kube-api-access-n5wnd\") pod \"console-f9d7485db-m2msx\" (UID: \"4ce6dbe8-0292-4172-81a3-f5cdfdba6120\") " pod="openshift-console/console-f9d7485db-m2msx" Dec 03 00:09:18 crc kubenswrapper[4953]: I1203 00:09:18.726545 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/b77462cc-f0ef-4108-b81f-0d6d412d7fa8-etcd-service-ca\") pod \"etcd-operator-b45778765-t968q\" (UID: \"b77462cc-f0ef-4108-b81f-0d6d412d7fa8\") " pod="openshift-etcd-operator/etcd-operator-b45778765-t968q" Dec 03 00:09:18 crc kubenswrapper[4953]: I1203 00:09:18.726572 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/4313395c-8501-4b5b-a4cb-2fc986fced03-ca-trust-extracted\") pod \"image-registry-697d97f7c8-t7ljb\" (UID: \"4313395c-8501-4b5b-a4cb-2fc986fced03\") " pod="openshift-image-registry/image-registry-697d97f7c8-t7ljb" Dec 03 00:09:18 crc kubenswrapper[4953]: I1203 00:09:18.726600 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fwqw4\" (UniqueName: \"kubernetes.io/projected/90247294-af78-4d04-8aed-af90f93a034a-kube-api-access-fwqw4\") pod \"route-controller-manager-6576b87f9c-52hrr\" (UID: \"90247294-af78-4d04-8aed-af90f93a034a\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-52hrr" Dec 03 00:09:18 crc kubenswrapper[4953]: I1203 00:09:18.726631 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/4313395c-8501-4b5b-a4cb-2fc986fced03-installation-pull-secrets\") pod \"image-registry-697d97f7c8-t7ljb\" (UID: \"4313395c-8501-4b5b-a4cb-2fc986fced03\") " pod="openshift-image-registry/image-registry-697d97f7c8-t7ljb" Dec 03 00:09:18 crc kubenswrapper[4953]: I1203 00:09:18.726676 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/4ce6dbe8-0292-4172-81a3-f5cdfdba6120-trusted-ca-bundle\") pod \"console-f9d7485db-m2msx\" (UID: \"4ce6dbe8-0292-4172-81a3-f5cdfdba6120\") " pod="openshift-console/console-f9d7485db-m2msx" Dec 03 00:09:18 crc kubenswrapper[4953]: I1203 00:09:18.726746 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/90247294-af78-4d04-8aed-af90f93a034a-client-ca\") pod \"route-controller-manager-6576b87f9c-52hrr\" (UID: \"90247294-af78-4d04-8aed-af90f93a034a\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-52hrr" Dec 03 00:09:18 crc kubenswrapper[4953]: I1203 00:09:18.726767 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/2c280414-49b4-489f-b74a-13f6715744a6-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-l29cz\" (UID: \"2c280414-49b4-489f-b74a-13f6715744a6\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-l29cz" Dec 03 00:09:18 crc kubenswrapper[4953]: I1203 00:09:18.726807 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2000e3b0-f8e7-4f0a-9f59-8c9d73324586-serving-cert\") pod \"openshift-config-operator-7777fb866f-l7m9z\" (UID: \"2000e3b0-f8e7-4f0a-9f59-8c9d73324586\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-l7m9z" Dec 03 00:09:18 crc kubenswrapper[4953]: I1203 00:09:18.726835 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/4313395c-8501-4b5b-a4cb-2fc986fced03-registry-tls\") pod \"image-registry-697d97f7c8-t7ljb\" (UID: \"4313395c-8501-4b5b-a4cb-2fc986fced03\") " pod="openshift-image-registry/image-registry-697d97f7c8-t7ljb" Dec 03 00:09:18 crc kubenswrapper[4953]: I1203 00:09:18.726853 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b77462cc-f0ef-4108-b81f-0d6d412d7fa8-config\") pod \"etcd-operator-b45778765-t968q\" (UID: \"b77462cc-f0ef-4108-b81f-0d6d412d7fa8\") " pod="openshift-etcd-operator/etcd-operator-b45778765-t968q" Dec 03 00:09:18 crc kubenswrapper[4953]: I1203 00:09:18.726900 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/4ce6dbe8-0292-4172-81a3-f5cdfdba6120-console-config\") pod \"console-f9d7485db-m2msx\" (UID: \"4ce6dbe8-0292-4172-81a3-f5cdfdba6120\") " pod="openshift-console/console-f9d7485db-m2msx" Dec 03 00:09:18 crc kubenswrapper[4953]: I1203 00:09:18.726917 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/536278ea-21b6-4193-ad9d-4dfb7ce6c194-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-dlqdp\" (UID: \"536278ea-21b6-4193-ad9d-4dfb7ce6c194\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-dlqdp" Dec 03 00:09:18 crc kubenswrapper[4953]: I1203 00:09:18.726947 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/4313395c-8501-4b5b-a4cb-2fc986fced03-registry-certificates\") pod \"image-registry-697d97f7c8-t7ljb\" (UID: \"4313395c-8501-4b5b-a4cb-2fc986fced03\") " pod="openshift-image-registry/image-registry-697d97f7c8-t7ljb" Dec 03 00:09:18 crc kubenswrapper[4953]: I1203 00:09:18.726964 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6d44369d-4150-4832-9098-88d3c4b1133e-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-gdstj\" (UID: \"6d44369d-4150-4832-9098-88d3c4b1133e\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-gdstj" Dec 03 00:09:18 crc kubenswrapper[4953]: I1203 00:09:18.726993 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/4ce6dbe8-0292-4172-81a3-f5cdfdba6120-oauth-serving-cert\") pod \"console-f9d7485db-m2msx\" (UID: \"4ce6dbe8-0292-4172-81a3-f5cdfdba6120\") " pod="openshift-console/console-f9d7485db-m2msx" Dec 03 00:09:18 crc kubenswrapper[4953]: I1203 00:09:18.727028 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/90247294-af78-4d04-8aed-af90f93a034a-config\") pod \"route-controller-manager-6576b87f9c-52hrr\" (UID: \"90247294-af78-4d04-8aed-af90f93a034a\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-52hrr" Dec 03 00:09:18 crc kubenswrapper[4953]: I1203 00:09:18.727045 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dcpj6\" (UniqueName: \"kubernetes.io/projected/6d44369d-4150-4832-9098-88d3c4b1133e-kube-api-access-dcpj6\") pod \"openshift-apiserver-operator-796bbdcf4f-gdstj\" (UID: \"6d44369d-4150-4832-9098-88d3c4b1133e\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-gdstj" Dec 03 00:09:18 crc kubenswrapper[4953]: I1203 00:09:18.727064 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b77462cc-f0ef-4108-b81f-0d6d412d7fa8-serving-cert\") pod \"etcd-operator-b45778765-t968q\" (UID: \"b77462cc-f0ef-4108-b81f-0d6d412d7fa8\") " pod="openshift-etcd-operator/etcd-operator-b45778765-t968q" Dec 03 00:09:18 crc kubenswrapper[4953]: I1203 00:09:18.727114 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h4grc\" (UniqueName: \"kubernetes.io/projected/a8a759e4-551c-41f9-bbab-9cfbf490777a-kube-api-access-h4grc\") pod \"machine-config-controller-84d6567774-95vxq\" (UID: \"a8a759e4-551c-41f9-bbab-9cfbf490777a\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-95vxq" Dec 03 00:09:18 crc kubenswrapper[4953]: I1203 00:09:18.727155 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-t7ljb\" (UID: \"4313395c-8501-4b5b-a4cb-2fc986fced03\") " pod="openshift-image-registry/image-registry-697d97f7c8-t7ljb" Dec 03 00:09:18 crc kubenswrapper[4953]: I1203 00:09:18.727175 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/a8a759e4-551c-41f9-bbab-9cfbf490777a-proxy-tls\") pod \"machine-config-controller-84d6567774-95vxq\" (UID: \"a8a759e4-551c-41f9-bbab-9cfbf490777a\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-95vxq" Dec 03 00:09:18 crc kubenswrapper[4953]: I1203 00:09:18.727192 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/90247294-af78-4d04-8aed-af90f93a034a-serving-cert\") pod \"route-controller-manager-6576b87f9c-52hrr\" (UID: \"90247294-af78-4d04-8aed-af90f93a034a\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-52hrr" Dec 03 00:09:18 crc kubenswrapper[4953]: I1203 00:09:18.727222 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c7jxx\" (UniqueName: \"kubernetes.io/projected/4313395c-8501-4b5b-a4cb-2fc986fced03-kube-api-access-c7jxx\") pod \"image-registry-697d97f7c8-t7ljb\" (UID: \"4313395c-8501-4b5b-a4cb-2fc986fced03\") " pod="openshift-image-registry/image-registry-697d97f7c8-t7ljb" Dec 03 00:09:18 crc kubenswrapper[4953]: I1203 00:09:18.727339 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2c280414-49b4-489f-b74a-13f6715744a6-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-l29cz\" (UID: \"2c280414-49b4-489f-b74a-13f6715744a6\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-l29cz" Dec 03 00:09:18 crc kubenswrapper[4953]: I1203 00:09:18.727420 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sbr4n\" (UniqueName: \"kubernetes.io/projected/2000e3b0-f8e7-4f0a-9f59-8c9d73324586-kube-api-access-sbr4n\") pod \"openshift-config-operator-7777fb866f-l7m9z\" (UID: \"2000e3b0-f8e7-4f0a-9f59-8c9d73324586\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-l7m9z" Dec 03 00:09:18 crc kubenswrapper[4953]: I1203 00:09:18.728254 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6d44369d-4150-4832-9098-88d3c4b1133e-config\") pod \"openshift-apiserver-operator-796bbdcf4f-gdstj\" (UID: \"6d44369d-4150-4832-9098-88d3c4b1133e\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-gdstj" Dec 03 00:09:18 crc kubenswrapper[4953]: E1203 00:09:18.728299 4953 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 00:09:19.228279118 +0000 UTC m=+148.212560081 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-t7ljb" (UID: "4313395c-8501-4b5b-a4cb-2fc986fced03") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 00:09:18 crc kubenswrapper[4953]: I1203 00:09:18.728339 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/b77462cc-f0ef-4108-b81f-0d6d412d7fa8-etcd-ca\") pod \"etcd-operator-b45778765-t968q\" (UID: \"b77462cc-f0ef-4108-b81f-0d6d412d7fa8\") " pod="openshift-etcd-operator/etcd-operator-b45778765-t968q" Dec 03 00:09:18 crc kubenswrapper[4953]: I1203 00:09:18.728363 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kcwc5\" (UniqueName: \"kubernetes.io/projected/536278ea-21b6-4193-ad9d-4dfb7ce6c194-kube-api-access-kcwc5\") pod \"package-server-manager-789f6589d5-dlqdp\" (UID: \"536278ea-21b6-4193-ad9d-4dfb7ce6c194\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-dlqdp" Dec 03 00:09:18 crc kubenswrapper[4953]: I1203 00:09:18.728559 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2c280414-49b4-489f-b74a-13f6715744a6-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-l29cz\" (UID: \"2c280414-49b4-489f-b74a-13f6715744a6\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-l29cz" Dec 03 00:09:18 crc kubenswrapper[4953]: I1203 00:09:18.728648 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-55l7h\" (UniqueName: \"kubernetes.io/projected/b77462cc-f0ef-4108-b81f-0d6d412d7fa8-kube-api-access-55l7h\") pod \"etcd-operator-b45778765-t968q\" (UID: \"b77462cc-f0ef-4108-b81f-0d6d412d7fa8\") " pod="openshift-etcd-operator/etcd-operator-b45778765-t968q" Dec 03 00:09:18 crc kubenswrapper[4953]: I1203 00:09:18.729167 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/2000e3b0-f8e7-4f0a-9f59-8c9d73324586-available-featuregates\") pod \"openshift-config-operator-7777fb866f-l7m9z\" (UID: \"2000e3b0-f8e7-4f0a-9f59-8c9d73324586\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-l7m9z" Dec 03 00:09:18 crc kubenswrapper[4953]: I1203 00:09:18.729999 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/4313395c-8501-4b5b-a4cb-2fc986fced03-trusted-ca\") pod \"image-registry-697d97f7c8-t7ljb\" (UID: \"4313395c-8501-4b5b-a4cb-2fc986fced03\") " pod="openshift-image-registry/image-registry-697d97f7c8-t7ljb" Dec 03 00:09:18 crc kubenswrapper[4953]: I1203 00:09:18.730090 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/a8a759e4-551c-41f9-bbab-9cfbf490777a-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-95vxq\" (UID: \"a8a759e4-551c-41f9-bbab-9cfbf490777a\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-95vxq" Dec 03 00:09:18 crc kubenswrapper[4953]: I1203 00:09:18.730592 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/b77462cc-f0ef-4108-b81f-0d6d412d7fa8-etcd-client\") pod \"etcd-operator-b45778765-t968q\" (UID: \"b77462cc-f0ef-4108-b81f-0d6d412d7fa8\") " pod="openshift-etcd-operator/etcd-operator-b45778765-t968q" Dec 03 00:09:18 crc kubenswrapper[4953]: I1203 00:09:18.741093 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-gsszn" Dec 03 00:09:18 crc kubenswrapper[4953]: I1203 00:09:18.819562 4953 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-2bh27"] Dec 03 00:09:18 crc kubenswrapper[4953]: I1203 00:09:18.832375 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 00:09:18 crc kubenswrapper[4953]: I1203 00:09:18.832596 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6j5fc\" (UniqueName: \"kubernetes.io/projected/7b91526b-2492-4c17-949b-0ba3f8c221ea-kube-api-access-6j5fc\") pod \"collect-profiles-29412000-mqgxv\" (UID: \"7b91526b-2492-4c17-949b-0ba3f8c221ea\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412000-mqgxv" Dec 03 00:09:18 crc kubenswrapper[4953]: I1203 00:09:18.832618 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/7124b81d-b87c-45de-a112-20602c4d9adb-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-kx9dt\" (UID: \"7124b81d-b87c-45de-a112-20602c4d9adb\") " pod="openshift-marketplace/marketplace-operator-79b997595-kx9dt" Dec 03 00:09:18 crc kubenswrapper[4953]: I1203 00:09:18.832639 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vzp6w\" (UniqueName: \"kubernetes.io/projected/04f69b6e-9193-4055-8e50-e87fa663ca05-kube-api-access-vzp6w\") pod \"machine-config-server-gpwzf\" (UID: \"04f69b6e-9193-4055-8e50-e87fa663ca05\") " pod="openshift-machine-config-operator/machine-config-server-gpwzf" Dec 03 00:09:18 crc kubenswrapper[4953]: I1203 00:09:18.832657 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/73033ee8-cae5-4acc-8c77-60905eed62f2-socket-dir\") pod \"csi-hostpathplugin-2xt2r\" (UID: \"73033ee8-cae5-4acc-8c77-60905eed62f2\") " pod="hostpath-provisioner/csi-hostpathplugin-2xt2r" Dec 03 00:09:18 crc kubenswrapper[4953]: I1203 00:09:18.832697 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1ffff94d-cd60-4079-86d6-ff10eb285ca2-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-z5glq\" (UID: \"1ffff94d-cd60-4079-86d6-ff10eb285ca2\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-z5glq" Dec 03 00:09:18 crc kubenswrapper[4953]: I1203 00:09:18.832717 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/174354b2-d5cc-495a-ac13-cb56923aca08-signing-key\") pod \"service-ca-9c57cc56f-zd9fs\" (UID: \"174354b2-d5cc-495a-ac13-cb56923aca08\") " pod="openshift-service-ca/service-ca-9c57cc56f-zd9fs" Dec 03 00:09:18 crc kubenswrapper[4953]: I1203 00:09:18.833006 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kcwc5\" (UniqueName: \"kubernetes.io/projected/536278ea-21b6-4193-ad9d-4dfb7ce6c194-kube-api-access-kcwc5\") pod \"package-server-manager-789f6589d5-dlqdp\" (UID: \"536278ea-21b6-4193-ad9d-4dfb7ce6c194\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-dlqdp" Dec 03 00:09:18 crc kubenswrapper[4953]: I1203 00:09:18.833052 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7cs9l\" (UniqueName: \"kubernetes.io/projected/e7e8210f-8915-49c3-8e95-f4427962f66a-kube-api-access-7cs9l\") pod \"olm-operator-6b444d44fb-h95rc\" (UID: \"e7e8210f-8915-49c3-8e95-f4427962f66a\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-h95rc" Dec 03 00:09:18 crc kubenswrapper[4953]: I1203 00:09:18.833071 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xb6nh\" (UniqueName: \"kubernetes.io/projected/1f9a783f-9a9e-47b6-bc01-9c25261b1f2b-kube-api-access-xb6nh\") pod \"dns-default-qbsjg\" (UID: \"1f9a783f-9a9e-47b6-bc01-9c25261b1f2b\") " pod="openshift-dns/dns-default-qbsjg" Dec 03 00:09:18 crc kubenswrapper[4953]: E1203 00:09:18.833135 4953 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 00:09:19.333089626 +0000 UTC m=+148.317370549 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 00:09:18 crc kubenswrapper[4953]: I1203 00:09:18.833236 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/4edc57f1-5c0c-4108-bb3c-b2f104c6e08b-service-ca-bundle\") pod \"router-default-5444994796-r8l2l\" (UID: \"4edc57f1-5c0c-4108-bb3c-b2f104c6e08b\") " pod="openshift-ingress/router-default-5444994796-r8l2l" Dec 03 00:09:18 crc kubenswrapper[4953]: I1203 00:09:18.833291 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-55l7h\" (UniqueName: \"kubernetes.io/projected/b77462cc-f0ef-4108-b81f-0d6d412d7fa8-kube-api-access-55l7h\") pod \"etcd-operator-b45778765-t968q\" (UID: \"b77462cc-f0ef-4108-b81f-0d6d412d7fa8\") " pod="openshift-etcd-operator/etcd-operator-b45778765-t968q" Dec 03 00:09:18 crc kubenswrapper[4953]: I1203 00:09:18.833334 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fc4d1764-e8db-415b-9650-ed9a03cc49d1-config\") pod \"kube-controller-manager-operator-78b949d7b-njtzc\" (UID: \"fc4d1764-e8db-415b-9650-ed9a03cc49d1\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-njtzc" Dec 03 00:09:18 crc kubenswrapper[4953]: I1203 00:09:18.833613 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/13c8b8e5-a48a-4376-be38-f4902b55ad94-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-jfp8j\" (UID: \"13c8b8e5-a48a-4376-be38-f4902b55ad94\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-jfp8j" Dec 03 00:09:18 crc kubenswrapper[4953]: I1203 00:09:18.833650 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/73033ee8-cae5-4acc-8c77-60905eed62f2-registration-dir\") pod \"csi-hostpathplugin-2xt2r\" (UID: \"73033ee8-cae5-4acc-8c77-60905eed62f2\") " pod="hostpath-provisioner/csi-hostpathplugin-2xt2r" Dec 03 00:09:18 crc kubenswrapper[4953]: I1203 00:09:18.833696 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/2000e3b0-f8e7-4f0a-9f59-8c9d73324586-available-featuregates\") pod \"openshift-config-operator-7777fb866f-l7m9z\" (UID: \"2000e3b0-f8e7-4f0a-9f59-8c9d73324586\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-l7m9z" Dec 03 00:09:18 crc kubenswrapper[4953]: I1203 00:09:18.833727 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sp7zr\" (UniqueName: \"kubernetes.io/projected/912368ff-455d-4a2d-a444-08c54448ca2a-kube-api-access-sp7zr\") pod \"migrator-59844c95c7-5fdxh\" (UID: \"912368ff-455d-4a2d-a444-08c54448ca2a\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-5fdxh" Dec 03 00:09:18 crc kubenswrapper[4953]: I1203 00:09:18.834156 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/2000e3b0-f8e7-4f0a-9f59-8c9d73324586-available-featuregates\") pod \"openshift-config-operator-7777fb866f-l7m9z\" (UID: \"2000e3b0-f8e7-4f0a-9f59-8c9d73324586\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-l7m9z" Dec 03 00:09:18 crc kubenswrapper[4953]: I1203 00:09:18.834236 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/4313395c-8501-4b5b-a4cb-2fc986fced03-trusted-ca\") pod \"image-registry-697d97f7c8-t7ljb\" (UID: \"4313395c-8501-4b5b-a4cb-2fc986fced03\") " pod="openshift-image-registry/image-registry-697d97f7c8-t7ljb" Dec 03 00:09:18 crc kubenswrapper[4953]: I1203 00:09:18.834287 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/e7e8210f-8915-49c3-8e95-f4427962f66a-profile-collector-cert\") pod \"olm-operator-6b444d44fb-h95rc\" (UID: \"e7e8210f-8915-49c3-8e95-f4427962f66a\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-h95rc" Dec 03 00:09:18 crc kubenswrapper[4953]: I1203 00:09:18.834436 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/b77462cc-f0ef-4108-b81f-0d6d412d7fa8-etcd-client\") pod \"etcd-operator-b45778765-t968q\" (UID: \"b77462cc-f0ef-4108-b81f-0d6d412d7fa8\") " pod="openshift-etcd-operator/etcd-operator-b45778765-t968q" Dec 03 00:09:18 crc kubenswrapper[4953]: I1203 00:09:18.834522 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/4da5668c-a644-4014-8358-790da0f313e2-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-xhc8f\" (UID: \"4da5668c-a644-4014-8358-790da0f313e2\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-xhc8f" Dec 03 00:09:18 crc kubenswrapper[4953]: I1203 00:09:18.835047 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/4ce6dbe8-0292-4172-81a3-f5cdfdba6120-console-oauth-config\") pod \"console-f9d7485db-m2msx\" (UID: \"4ce6dbe8-0292-4172-81a3-f5cdfdba6120\") " pod="openshift-console/console-f9d7485db-m2msx" Dec 03 00:09:18 crc kubenswrapper[4953]: I1203 00:09:18.835790 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fwqw4\" (UniqueName: \"kubernetes.io/projected/90247294-af78-4d04-8aed-af90f93a034a-kube-api-access-fwqw4\") pod \"route-controller-manager-6576b87f9c-52hrr\" (UID: \"90247294-af78-4d04-8aed-af90f93a034a\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-52hrr" Dec 03 00:09:18 crc kubenswrapper[4953]: I1203 00:09:18.835872 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/4edc57f1-5c0c-4108-bb3c-b2f104c6e08b-metrics-certs\") pod \"router-default-5444994796-r8l2l\" (UID: \"4edc57f1-5c0c-4108-bb3c-b2f104c6e08b\") " pod="openshift-ingress/router-default-5444994796-r8l2l" Dec 03 00:09:18 crc kubenswrapper[4953]: I1203 00:09:18.836071 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/4313395c-8501-4b5b-a4cb-2fc986fced03-installation-pull-secrets\") pod \"image-registry-697d97f7c8-t7ljb\" (UID: \"4313395c-8501-4b5b-a4cb-2fc986fced03\") " pod="openshift-image-registry/image-registry-697d97f7c8-t7ljb" Dec 03 00:09:18 crc kubenswrapper[4953]: I1203 00:09:18.836171 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wgsd8\" (UniqueName: \"kubernetes.io/projected/4da5668c-a644-4014-8358-790da0f313e2-kube-api-access-wgsd8\") pod \"control-plane-machine-set-operator-78cbb6b69f-xhc8f\" (UID: \"4da5668c-a644-4014-8358-790da0f313e2\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-xhc8f" Dec 03 00:09:18 crc kubenswrapper[4953]: I1203 00:09:18.836206 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kkx26\" (UniqueName: \"kubernetes.io/projected/174354b2-d5cc-495a-ac13-cb56923aca08-kube-api-access-kkx26\") pod \"service-ca-9c57cc56f-zd9fs\" (UID: \"174354b2-d5cc-495a-ac13-cb56923aca08\") " pod="openshift-service-ca/service-ca-9c57cc56f-zd9fs" Dec 03 00:09:18 crc kubenswrapper[4953]: I1203 00:09:18.836303 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/2c280414-49b4-489f-b74a-13f6715744a6-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-l29cz\" (UID: \"2c280414-49b4-489f-b74a-13f6715744a6\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-l29cz" Dec 03 00:09:18 crc kubenswrapper[4953]: I1203 00:09:18.836336 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/1ee4772d-f4e3-4da8-a520-805dc2317fdd-profile-collector-cert\") pod \"catalog-operator-68c6474976-brcxn\" (UID: \"1ee4772d-f4e3-4da8-a520-805dc2317fdd\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-brcxn" Dec 03 00:09:18 crc kubenswrapper[4953]: I1203 00:09:18.836408 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/7124b81d-b87c-45de-a112-20602c4d9adb-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-kx9dt\" (UID: \"7124b81d-b87c-45de-a112-20602c4d9adb\") " pod="openshift-marketplace/marketplace-operator-79b997595-kx9dt" Dec 03 00:09:18 crc kubenswrapper[4953]: I1203 00:09:18.836461 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/4313395c-8501-4b5b-a4cb-2fc986fced03-registry-tls\") pod \"image-registry-697d97f7c8-t7ljb\" (UID: \"4313395c-8501-4b5b-a4cb-2fc986fced03\") " pod="openshift-image-registry/image-registry-697d97f7c8-t7ljb" Dec 03 00:09:18 crc kubenswrapper[4953]: I1203 00:09:18.836511 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/4313395c-8501-4b5b-a4cb-2fc986fced03-trusted-ca\") pod \"image-registry-697d97f7c8-t7ljb\" (UID: \"4313395c-8501-4b5b-a4cb-2fc986fced03\") " pod="openshift-image-registry/image-registry-697d97f7c8-t7ljb" Dec 03 00:09:18 crc kubenswrapper[4953]: I1203 00:09:18.836532 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b77462cc-f0ef-4108-b81f-0d6d412d7fa8-config\") pod \"etcd-operator-b45778765-t968q\" (UID: \"b77462cc-f0ef-4108-b81f-0d6d412d7fa8\") " pod="openshift-etcd-operator/etcd-operator-b45778765-t968q" Dec 03 00:09:18 crc kubenswrapper[4953]: I1203 00:09:18.836645 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/48b0efc4-5e8a-4bd9-9016-26bdd4761bf1-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-kz9m9\" (UID: \"48b0efc4-5e8a-4bd9-9016-26bdd4761bf1\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-kz9m9" Dec 03 00:09:18 crc kubenswrapper[4953]: I1203 00:09:18.836702 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/4ce6dbe8-0292-4172-81a3-f5cdfdba6120-console-config\") pod \"console-f9d7485db-m2msx\" (UID: \"4ce6dbe8-0292-4172-81a3-f5cdfdba6120\") " pod="openshift-console/console-f9d7485db-m2msx" Dec 03 00:09:18 crc kubenswrapper[4953]: I1203 00:09:18.836729 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/536278ea-21b6-4193-ad9d-4dfb7ce6c194-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-dlqdp\" (UID: \"536278ea-21b6-4193-ad9d-4dfb7ce6c194\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-dlqdp" Dec 03 00:09:18 crc kubenswrapper[4953]: I1203 00:09:18.836825 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/4ce6dbe8-0292-4172-81a3-f5cdfdba6120-oauth-serving-cert\") pod \"console-f9d7485db-m2msx\" (UID: \"4ce6dbe8-0292-4172-81a3-f5cdfdba6120\") " pod="openshift-console/console-f9d7485db-m2msx" Dec 03 00:09:18 crc kubenswrapper[4953]: I1203 00:09:18.836853 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/13c8b8e5-a48a-4376-be38-f4902b55ad94-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-jfp8j\" (UID: \"13c8b8e5-a48a-4376-be38-f4902b55ad94\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-jfp8j" Dec 03 00:09:18 crc kubenswrapper[4953]: I1203 00:09:18.836895 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/90247294-af78-4d04-8aed-af90f93a034a-config\") pod \"route-controller-manager-6576b87f9c-52hrr\" (UID: \"90247294-af78-4d04-8aed-af90f93a034a\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-52hrr" Dec 03 00:09:18 crc kubenswrapper[4953]: I1203 00:09:18.837028 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dcpj6\" (UniqueName: \"kubernetes.io/projected/6d44369d-4150-4832-9098-88d3c4b1133e-kube-api-access-dcpj6\") pod \"openshift-apiserver-operator-796bbdcf4f-gdstj\" (UID: \"6d44369d-4150-4832-9098-88d3c4b1133e\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-gdstj" Dec 03 00:09:18 crc kubenswrapper[4953]: I1203 00:09:18.837060 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/73033ee8-cae5-4acc-8c77-60905eed62f2-csi-data-dir\") pod \"csi-hostpathplugin-2xt2r\" (UID: \"73033ee8-cae5-4acc-8c77-60905eed62f2\") " pod="hostpath-provisioner/csi-hostpathplugin-2xt2r" Dec 03 00:09:18 crc kubenswrapper[4953]: I1203 00:09:18.837303 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/c893caae-db46-4210-8be1-0ccf40bc9913-apiservice-cert\") pod \"packageserver-d55dfcdfc-qv29t\" (UID: \"c893caae-db46-4210-8be1-0ccf40bc9913\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-qv29t" Dec 03 00:09:18 crc kubenswrapper[4953]: I1203 00:09:18.837336 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1f9a783f-9a9e-47b6-bc01-9c25261b1f2b-config-volume\") pod \"dns-default-qbsjg\" (UID: \"1f9a783f-9a9e-47b6-bc01-9c25261b1f2b\") " pod="openshift-dns/dns-default-qbsjg" Dec 03 00:09:18 crc kubenswrapper[4953]: I1203 00:09:18.837443 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1ffff94d-cd60-4079-86d6-ff10eb285ca2-config\") pod \"kube-apiserver-operator-766d6c64bb-z5glq\" (UID: \"1ffff94d-cd60-4079-86d6-ff10eb285ca2\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-z5glq" Dec 03 00:09:18 crc kubenswrapper[4953]: I1203 00:09:18.837472 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c7jxx\" (UniqueName: \"kubernetes.io/projected/4313395c-8501-4b5b-a4cb-2fc986fced03-kube-api-access-c7jxx\") pod \"image-registry-697d97f7c8-t7ljb\" (UID: \"4313395c-8501-4b5b-a4cb-2fc986fced03\") " pod="openshift-image-registry/image-registry-697d97f7c8-t7ljb" Dec 03 00:09:18 crc kubenswrapper[4953]: I1203 00:09:18.837539 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5h6kc\" (UniqueName: \"kubernetes.io/projected/48b0efc4-5e8a-4bd9-9016-26bdd4761bf1-kube-api-access-5h6kc\") pod \"openshift-controller-manager-operator-756b6f6bc6-kz9m9\" (UID: \"48b0efc4-5e8a-4bd9-9016-26bdd4761bf1\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-kz9m9" Dec 03 00:09:18 crc kubenswrapper[4953]: I1203 00:09:18.837573 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2c280414-49b4-489f-b74a-13f6715744a6-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-l29cz\" (UID: \"2c280414-49b4-489f-b74a-13f6715744a6\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-l29cz" Dec 03 00:09:18 crc kubenswrapper[4953]: I1203 00:09:18.837599 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sbr4n\" (UniqueName: \"kubernetes.io/projected/2000e3b0-f8e7-4f0a-9f59-8c9d73324586-kube-api-access-sbr4n\") pod \"openshift-config-operator-7777fb866f-l7m9z\" (UID: \"2000e3b0-f8e7-4f0a-9f59-8c9d73324586\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-l7m9z" Dec 03 00:09:18 crc kubenswrapper[4953]: I1203 00:09:18.837622 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/73033ee8-cae5-4acc-8c77-60905eed62f2-plugins-dir\") pod \"csi-hostpathplugin-2xt2r\" (UID: \"73033ee8-cae5-4acc-8c77-60905eed62f2\") " pod="hostpath-provisioner/csi-hostpathplugin-2xt2r" Dec 03 00:09:18 crc kubenswrapper[4953]: I1203 00:09:18.837647 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/4edc57f1-5c0c-4108-bb3c-b2f104c6e08b-default-certificate\") pod \"router-default-5444994796-r8l2l\" (UID: \"4edc57f1-5c0c-4108-bb3c-b2f104c6e08b\") " pod="openshift-ingress/router-default-5444994796-r8l2l" Dec 03 00:09:18 crc kubenswrapper[4953]: I1203 00:09:18.837673 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6d44369d-4150-4832-9098-88d3c4b1133e-config\") pod \"openshift-apiserver-operator-796bbdcf4f-gdstj\" (UID: \"6d44369d-4150-4832-9098-88d3c4b1133e\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-gdstj" Dec 03 00:09:18 crc kubenswrapper[4953]: I1203 00:09:18.837697 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/b77462cc-f0ef-4108-b81f-0d6d412d7fa8-etcd-ca\") pod \"etcd-operator-b45778765-t968q\" (UID: \"b77462cc-f0ef-4108-b81f-0d6d412d7fa8\") " pod="openshift-etcd-operator/etcd-operator-b45778765-t968q" Dec 03 00:09:18 crc kubenswrapper[4953]: I1203 00:09:18.837722 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/c893caae-db46-4210-8be1-0ccf40bc9913-webhook-cert\") pod \"packageserver-d55dfcdfc-qv29t\" (UID: \"c893caae-db46-4210-8be1-0ccf40bc9913\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-qv29t" Dec 03 00:09:18 crc kubenswrapper[4953]: I1203 00:09:18.837773 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2c280414-49b4-489f-b74a-13f6715744a6-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-l29cz\" (UID: \"2c280414-49b4-489f-b74a-13f6715744a6\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-l29cz" Dec 03 00:09:18 crc kubenswrapper[4953]: I1203 00:09:18.837795 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/e7e8210f-8915-49c3-8e95-f4427962f66a-srv-cert\") pod \"olm-operator-6b444d44fb-h95rc\" (UID: \"e7e8210f-8915-49c3-8e95-f4427962f66a\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-h95rc" Dec 03 00:09:18 crc kubenswrapper[4953]: I1203 00:09:18.837817 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/174354b2-d5cc-495a-ac13-cb56923aca08-signing-cabundle\") pod \"service-ca-9c57cc56f-zd9fs\" (UID: \"174354b2-d5cc-495a-ac13-cb56923aca08\") " pod="openshift-service-ca/service-ca-9c57cc56f-zd9fs" Dec 03 00:09:18 crc kubenswrapper[4953]: I1203 00:09:18.837851 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/fc4d1764-e8db-415b-9650-ed9a03cc49d1-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-njtzc\" (UID: \"fc4d1764-e8db-415b-9650-ed9a03cc49d1\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-njtzc" Dec 03 00:09:18 crc kubenswrapper[4953]: I1203 00:09:18.837887 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6whsj\" (UniqueName: \"kubernetes.io/projected/73033ee8-cae5-4acc-8c77-60905eed62f2-kube-api-access-6whsj\") pod \"csi-hostpathplugin-2xt2r\" (UID: \"73033ee8-cae5-4acc-8c77-60905eed62f2\") " pod="hostpath-provisioner/csi-hostpathplugin-2xt2r" Dec 03 00:09:18 crc kubenswrapper[4953]: I1203 00:09:18.837907 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/16c23c64-3f6c-4ce9-9118-cfeb890dfeed-config\") pod \"service-ca-operator-777779d784-fcw6n\" (UID: \"16c23c64-3f6c-4ce9-9118-cfeb890dfeed\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-fcw6n" Dec 03 00:09:18 crc kubenswrapper[4953]: I1203 00:09:18.839109 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/90247294-af78-4d04-8aed-af90f93a034a-config\") pod \"route-controller-manager-6576b87f9c-52hrr\" (UID: \"90247294-af78-4d04-8aed-af90f93a034a\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-52hrr" Dec 03 00:09:18 crc kubenswrapper[4953]: I1203 00:09:18.840634 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6d44369d-4150-4832-9098-88d3c4b1133e-config\") pod \"openshift-apiserver-operator-796bbdcf4f-gdstj\" (UID: \"6d44369d-4150-4832-9098-88d3c4b1133e\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-gdstj" Dec 03 00:09:18 crc kubenswrapper[4953]: I1203 00:09:18.840638 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b77462cc-f0ef-4108-b81f-0d6d412d7fa8-config\") pod \"etcd-operator-b45778765-t968q\" (UID: \"b77462cc-f0ef-4108-b81f-0d6d412d7fa8\") " pod="openshift-etcd-operator/etcd-operator-b45778765-t968q" Dec 03 00:09:18 crc kubenswrapper[4953]: I1203 00:09:18.840922 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/4ce6dbe8-0292-4172-81a3-f5cdfdba6120-console-oauth-config\") pod \"console-f9d7485db-m2msx\" (UID: \"4ce6dbe8-0292-4172-81a3-f5cdfdba6120\") " pod="openshift-console/console-f9d7485db-m2msx" Dec 03 00:09:18 crc kubenswrapper[4953]: I1203 00:09:18.846731 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/4ce6dbe8-0292-4172-81a3-f5cdfdba6120-oauth-serving-cert\") pod \"console-f9d7485db-m2msx\" (UID: \"4ce6dbe8-0292-4172-81a3-f5cdfdba6120\") " pod="openshift-console/console-f9d7485db-m2msx" Dec 03 00:09:18 crc kubenswrapper[4953]: I1203 00:09:18.847120 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2c280414-49b4-489f-b74a-13f6715744a6-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-l29cz\" (UID: \"2c280414-49b4-489f-b74a-13f6715744a6\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-l29cz" Dec 03 00:09:18 crc kubenswrapper[4953]: I1203 00:09:18.837966 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/a8a759e4-551c-41f9-bbab-9cfbf490777a-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-95vxq\" (UID: \"a8a759e4-551c-41f9-bbab-9cfbf490777a\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-95vxq" Dec 03 00:09:18 crc kubenswrapper[4953]: I1203 00:09:18.847364 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/b77462cc-f0ef-4108-b81f-0d6d412d7fa8-etcd-client\") pod \"etcd-operator-b45778765-t968q\" (UID: \"b77462cc-f0ef-4108-b81f-0d6d412d7fa8\") " pod="openshift-etcd-operator/etcd-operator-b45778765-t968q" Dec 03 00:09:18 crc kubenswrapper[4953]: I1203 00:09:18.847830 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/4ce6dbe8-0292-4172-81a3-f5cdfdba6120-console-config\") pod \"console-f9d7485db-m2msx\" (UID: \"4ce6dbe8-0292-4172-81a3-f5cdfdba6120\") " pod="openshift-console/console-f9d7485db-m2msx" Dec 03 00:09:18 crc kubenswrapper[4953]: I1203 00:09:18.847884 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/4313395c-8501-4b5b-a4cb-2fc986fced03-registry-tls\") pod \"image-registry-697d97f7c8-t7ljb\" (UID: \"4313395c-8501-4b5b-a4cb-2fc986fced03\") " pod="openshift-image-registry/image-registry-697d97f7c8-t7ljb" Dec 03 00:09:18 crc kubenswrapper[4953]: I1203 00:09:18.848672 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/4313395c-8501-4b5b-a4cb-2fc986fced03-installation-pull-secrets\") pod \"image-registry-697d97f7c8-t7ljb\" (UID: \"4313395c-8501-4b5b-a4cb-2fc986fced03\") " pod="openshift-image-registry/image-registry-697d97f7c8-t7ljb" Dec 03 00:09:18 crc kubenswrapper[4953]: I1203 00:09:18.850270 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/b77462cc-f0ef-4108-b81f-0d6d412d7fa8-etcd-ca\") pod \"etcd-operator-b45778765-t968q\" (UID: \"b77462cc-f0ef-4108-b81f-0d6d412d7fa8\") " pod="openshift-etcd-operator/etcd-operator-b45778765-t968q" Dec 03 00:09:18 crc kubenswrapper[4953]: I1203 00:09:18.854420 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7gsv5\" (UniqueName: \"kubernetes.io/projected/13c8b8e5-a48a-4376-be38-f4902b55ad94-kube-api-access-7gsv5\") pod \"kube-storage-version-migrator-operator-b67b599dd-jfp8j\" (UID: \"13c8b8e5-a48a-4376-be38-f4902b55ad94\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-jfp8j" Dec 03 00:09:18 crc kubenswrapper[4953]: I1203 00:09:18.854618 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1ffff94d-cd60-4079-86d6-ff10eb285ca2-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-z5glq\" (UID: \"1ffff94d-cd60-4079-86d6-ff10eb285ca2\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-z5glq" Dec 03 00:09:18 crc kubenswrapper[4953]: I1203 00:09:18.855571 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5bm5b\" (UniqueName: \"kubernetes.io/projected/7124b81d-b87c-45de-a112-20602c4d9adb-kube-api-access-5bm5b\") pod \"marketplace-operator-79b997595-kx9dt\" (UID: \"7124b81d-b87c-45de-a112-20602c4d9adb\") " pod="openshift-marketplace/marketplace-operator-79b997595-kx9dt" Dec 03 00:09:18 crc kubenswrapper[4953]: I1203 00:09:18.855824 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/4313395c-8501-4b5b-a4cb-2fc986fced03-bound-sa-token\") pod \"image-registry-697d97f7c8-t7ljb\" (UID: \"4313395c-8501-4b5b-a4cb-2fc986fced03\") " pod="openshift-image-registry/image-registry-697d97f7c8-t7ljb" Dec 03 00:09:18 crc kubenswrapper[4953]: I1203 00:09:18.855863 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/4ce6dbe8-0292-4172-81a3-f5cdfdba6120-console-serving-cert\") pod \"console-f9d7485db-m2msx\" (UID: \"4ce6dbe8-0292-4172-81a3-f5cdfdba6120\") " pod="openshift-console/console-f9d7485db-m2msx" Dec 03 00:09:18 crc kubenswrapper[4953]: I1203 00:09:18.856177 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2c280414-49b4-489f-b74a-13f6715744a6-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-l29cz\" (UID: \"2c280414-49b4-489f-b74a-13f6715744a6\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-l29cz" Dec 03 00:09:18 crc kubenswrapper[4953]: I1203 00:09:18.858074 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l7g4k\" (UniqueName: \"kubernetes.io/projected/1ee4772d-f4e3-4da8-a520-805dc2317fdd-kube-api-access-l7g4k\") pod \"catalog-operator-68c6474976-brcxn\" (UID: \"1ee4772d-f4e3-4da8-a520-805dc2317fdd\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-brcxn" Dec 03 00:09:18 crc kubenswrapper[4953]: I1203 00:09:18.860006 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/a8a759e4-551c-41f9-bbab-9cfbf490777a-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-95vxq\" (UID: \"a8a759e4-551c-41f9-bbab-9cfbf490777a\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-95vxq" Dec 03 00:09:18 crc kubenswrapper[4953]: I1203 00:09:18.861359 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/4ce6dbe8-0292-4172-81a3-f5cdfdba6120-service-ca\") pod \"console-f9d7485db-m2msx\" (UID: \"4ce6dbe8-0292-4172-81a3-f5cdfdba6120\") " pod="openshift-console/console-f9d7485db-m2msx" Dec 03 00:09:18 crc kubenswrapper[4953]: I1203 00:09:18.861511 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n5wnd\" (UniqueName: \"kubernetes.io/projected/4ce6dbe8-0292-4172-81a3-f5cdfdba6120-kube-api-access-n5wnd\") pod \"console-f9d7485db-m2msx\" (UID: \"4ce6dbe8-0292-4172-81a3-f5cdfdba6120\") " pod="openshift-console/console-f9d7485db-m2msx" Dec 03 00:09:18 crc kubenswrapper[4953]: I1203 00:09:18.861718 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/b77462cc-f0ef-4108-b81f-0d6d412d7fa8-etcd-service-ca\") pod \"etcd-operator-b45778765-t968q\" (UID: \"b77462cc-f0ef-4108-b81f-0d6d412d7fa8\") " pod="openshift-etcd-operator/etcd-operator-b45778765-t968q" Dec 03 00:09:18 crc kubenswrapper[4953]: I1203 00:09:18.861860 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/1f9a783f-9a9e-47b6-bc01-9c25261b1f2b-metrics-tls\") pod \"dns-default-qbsjg\" (UID: \"1f9a783f-9a9e-47b6-bc01-9c25261b1f2b\") " pod="openshift-dns/dns-default-qbsjg" Dec 03 00:09:18 crc kubenswrapper[4953]: I1203 00:09:18.862000 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/4edc57f1-5c0c-4108-bb3c-b2f104c6e08b-stats-auth\") pod \"router-default-5444994796-r8l2l\" (UID: \"4edc57f1-5c0c-4108-bb3c-b2f104c6e08b\") " pod="openshift-ingress/router-default-5444994796-r8l2l" Dec 03 00:09:18 crc kubenswrapper[4953]: I1203 00:09:18.862681 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/b77462cc-f0ef-4108-b81f-0d6d412d7fa8-etcd-service-ca\") pod \"etcd-operator-b45778765-t968q\" (UID: \"b77462cc-f0ef-4108-b81f-0d6d412d7fa8\") " pod="openshift-etcd-operator/etcd-operator-b45778765-t968q" Dec 03 00:09:18 crc kubenswrapper[4953]: I1203 00:09:18.862762 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-28pml\" (UniqueName: \"kubernetes.io/projected/16c23c64-3f6c-4ce9-9118-cfeb890dfeed-kube-api-access-28pml\") pod \"service-ca-operator-777779d784-fcw6n\" (UID: \"16c23c64-3f6c-4ce9-9118-cfeb890dfeed\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-fcw6n" Dec 03 00:09:18 crc kubenswrapper[4953]: I1203 00:09:18.862837 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/4313395c-8501-4b5b-a4cb-2fc986fced03-ca-trust-extracted\") pod \"image-registry-697d97f7c8-t7ljb\" (UID: \"4313395c-8501-4b5b-a4cb-2fc986fced03\") " pod="openshift-image-registry/image-registry-697d97f7c8-t7ljb" Dec 03 00:09:18 crc kubenswrapper[4953]: I1203 00:09:18.862915 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/16c23c64-3f6c-4ce9-9118-cfeb890dfeed-serving-cert\") pod \"service-ca-operator-777779d784-fcw6n\" (UID: \"16c23c64-3f6c-4ce9-9118-cfeb890dfeed\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-fcw6n" Dec 03 00:09:18 crc kubenswrapper[4953]: I1203 00:09:18.863040 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/400790d0-9d4f-4698-bd68-8085ac29ec9c-cert\") pod \"ingress-canary-llvsw\" (UID: \"400790d0-9d4f-4698-bd68-8085ac29ec9c\") " pod="openshift-ingress-canary/ingress-canary-llvsw" Dec 03 00:09:18 crc kubenswrapper[4953]: I1203 00:09:18.863074 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/04f69b6e-9193-4055-8e50-e87fa663ca05-node-bootstrap-token\") pod \"machine-config-server-gpwzf\" (UID: \"04f69b6e-9193-4055-8e50-e87fa663ca05\") " pod="openshift-machine-config-operator/machine-config-server-gpwzf" Dec 03 00:09:18 crc kubenswrapper[4953]: I1203 00:09:18.863271 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/4ce6dbe8-0292-4172-81a3-f5cdfdba6120-trusted-ca-bundle\") pod \"console-f9d7485db-m2msx\" (UID: \"4ce6dbe8-0292-4172-81a3-f5cdfdba6120\") " pod="openshift-console/console-f9d7485db-m2msx" Dec 03 00:09:18 crc kubenswrapper[4953]: I1203 00:09:18.863386 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/4313395c-8501-4b5b-a4cb-2fc986fced03-ca-trust-extracted\") pod \"image-registry-697d97f7c8-t7ljb\" (UID: \"4313395c-8501-4b5b-a4cb-2fc986fced03\") " pod="openshift-image-registry/image-registry-697d97f7c8-t7ljb" Dec 03 00:09:18 crc kubenswrapper[4953]: I1203 00:09:18.863427 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/90247294-af78-4d04-8aed-af90f93a034a-client-ca\") pod \"route-controller-manager-6576b87f9c-52hrr\" (UID: \"90247294-af78-4d04-8aed-af90f93a034a\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-52hrr" Dec 03 00:09:18 crc kubenswrapper[4953]: I1203 00:09:18.863685 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2000e3b0-f8e7-4f0a-9f59-8c9d73324586-serving-cert\") pod \"openshift-config-operator-7777fb866f-l7m9z\" (UID: \"2000e3b0-f8e7-4f0a-9f59-8c9d73324586\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-l7m9z" Dec 03 00:09:18 crc kubenswrapper[4953]: I1203 00:09:18.865107 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/7b91526b-2492-4c17-949b-0ba3f8c221ea-config-volume\") pod \"collect-profiles-29412000-mqgxv\" (UID: \"7b91526b-2492-4c17-949b-0ba3f8c221ea\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412000-mqgxv" Dec 03 00:09:18 crc kubenswrapper[4953]: I1203 00:09:18.866570 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/4ce6dbe8-0292-4172-81a3-f5cdfdba6120-service-ca\") pod \"console-f9d7485db-m2msx\" (UID: \"4ce6dbe8-0292-4172-81a3-f5cdfdba6120\") " pod="openshift-console/console-f9d7485db-m2msx" Dec 03 00:09:18 crc kubenswrapper[4953]: I1203 00:09:18.868991 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/90247294-af78-4d04-8aed-af90f93a034a-client-ca\") pod \"route-controller-manager-6576b87f9c-52hrr\" (UID: \"90247294-af78-4d04-8aed-af90f93a034a\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-52hrr" Dec 03 00:09:18 crc kubenswrapper[4953]: I1203 00:09:18.869526 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/536278ea-21b6-4193-ad9d-4dfb7ce6c194-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-dlqdp\" (UID: \"536278ea-21b6-4193-ad9d-4dfb7ce6c194\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-dlqdp" Dec 03 00:09:18 crc kubenswrapper[4953]: I1203 00:09:18.876591 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2000e3b0-f8e7-4f0a-9f59-8c9d73324586-serving-cert\") pod \"openshift-config-operator-7777fb866f-l7m9z\" (UID: \"2000e3b0-f8e7-4f0a-9f59-8c9d73324586\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-l7m9z" Dec 03 00:09:18 crc kubenswrapper[4953]: I1203 00:09:18.899262 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/4ce6dbe8-0292-4172-81a3-f5cdfdba6120-trusted-ca-bundle\") pod \"console-f9d7485db-m2msx\" (UID: \"4ce6dbe8-0292-4172-81a3-f5cdfdba6120\") " pod="openshift-console/console-f9d7485db-m2msx" Dec 03 00:09:18 crc kubenswrapper[4953]: I1203 00:09:18.899519 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/4ce6dbe8-0292-4172-81a3-f5cdfdba6120-console-serving-cert\") pod \"console-f9d7485db-m2msx\" (UID: \"4ce6dbe8-0292-4172-81a3-f5cdfdba6120\") " pod="openshift-console/console-f9d7485db-m2msx" Dec 03 00:09:18 crc kubenswrapper[4953]: I1203 00:09:18.902413 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dvf62\" (UniqueName: \"kubernetes.io/projected/c893caae-db46-4210-8be1-0ccf40bc9913-kube-api-access-dvf62\") pod \"packageserver-d55dfcdfc-qv29t\" (UID: \"c893caae-db46-4210-8be1-0ccf40bc9913\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-qv29t" Dec 03 00:09:18 crc kubenswrapper[4953]: I1203 00:09:18.902465 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/7b91526b-2492-4c17-949b-0ba3f8c221ea-secret-volume\") pod \"collect-profiles-29412000-mqgxv\" (UID: \"7b91526b-2492-4c17-949b-0ba3f8c221ea\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412000-mqgxv" Dec 03 00:09:18 crc kubenswrapper[4953]: I1203 00:09:18.902513 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/4313395c-8501-4b5b-a4cb-2fc986fced03-registry-certificates\") pod \"image-registry-697d97f7c8-t7ljb\" (UID: \"4313395c-8501-4b5b-a4cb-2fc986fced03\") " pod="openshift-image-registry/image-registry-697d97f7c8-t7ljb" Dec 03 00:09:18 crc kubenswrapper[4953]: I1203 00:09:18.902541 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6d44369d-4150-4832-9098-88d3c4b1133e-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-gdstj\" (UID: \"6d44369d-4150-4832-9098-88d3c4b1133e\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-gdstj" Dec 03 00:09:18 crc kubenswrapper[4953]: I1203 00:09:18.902564 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/1ee4772d-f4e3-4da8-a520-805dc2317fdd-srv-cert\") pod \"catalog-operator-68c6474976-brcxn\" (UID: \"1ee4772d-f4e3-4da8-a520-805dc2317fdd\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-brcxn" Dec 03 00:09:18 crc kubenswrapper[4953]: I1203 00:09:18.902594 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t4dws\" (UniqueName: \"kubernetes.io/projected/400790d0-9d4f-4698-bd68-8085ac29ec9c-kube-api-access-t4dws\") pod \"ingress-canary-llvsw\" (UID: \"400790d0-9d4f-4698-bd68-8085ac29ec9c\") " pod="openshift-ingress-canary/ingress-canary-llvsw" Dec 03 00:09:18 crc kubenswrapper[4953]: I1203 00:09:18.903148 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b77462cc-f0ef-4108-b81f-0d6d412d7fa8-serving-cert\") pod \"etcd-operator-b45778765-t968q\" (UID: \"b77462cc-f0ef-4108-b81f-0d6d412d7fa8\") " pod="openshift-etcd-operator/etcd-operator-b45778765-t968q" Dec 03 00:09:18 crc kubenswrapper[4953]: I1203 00:09:18.905983 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h4grc\" (UniqueName: \"kubernetes.io/projected/a8a759e4-551c-41f9-bbab-9cfbf490777a-kube-api-access-h4grc\") pod \"machine-config-controller-84d6567774-95vxq\" (UID: \"a8a759e4-551c-41f9-bbab-9cfbf490777a\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-95vxq" Dec 03 00:09:18 crc kubenswrapper[4953]: I1203 00:09:18.906029 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fc4d1764-e8db-415b-9650-ed9a03cc49d1-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-njtzc\" (UID: \"fc4d1764-e8db-415b-9650-ed9a03cc49d1\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-njtzc" Dec 03 00:09:18 crc kubenswrapper[4953]: I1203 00:09:18.906053 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/04f69b6e-9193-4055-8e50-e87fa663ca05-certs\") pod \"machine-config-server-gpwzf\" (UID: \"04f69b6e-9193-4055-8e50-e87fa663ca05\") " pod="openshift-machine-config-operator/machine-config-server-gpwzf" Dec 03 00:09:18 crc kubenswrapper[4953]: I1203 00:09:18.906092 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-t7ljb\" (UID: \"4313395c-8501-4b5b-a4cb-2fc986fced03\") " pod="openshift-image-registry/image-registry-697d97f7c8-t7ljb" Dec 03 00:09:18 crc kubenswrapper[4953]: I1203 00:09:18.906117 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/c893caae-db46-4210-8be1-0ccf40bc9913-tmpfs\") pod \"packageserver-d55dfcdfc-qv29t\" (UID: \"c893caae-db46-4210-8be1-0ccf40bc9913\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-qv29t" Dec 03 00:09:18 crc kubenswrapper[4953]: I1203 00:09:18.906159 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/a8a759e4-551c-41f9-bbab-9cfbf490777a-proxy-tls\") pod \"machine-config-controller-84d6567774-95vxq\" (UID: \"a8a759e4-551c-41f9-bbab-9cfbf490777a\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-95vxq" Dec 03 00:09:18 crc kubenswrapper[4953]: I1203 00:09:18.906189 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/90247294-af78-4d04-8aed-af90f93a034a-serving-cert\") pod \"route-controller-manager-6576b87f9c-52hrr\" (UID: \"90247294-af78-4d04-8aed-af90f93a034a\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-52hrr" Dec 03 00:09:18 crc kubenswrapper[4953]: I1203 00:09:18.906215 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/48b0efc4-5e8a-4bd9-9016-26bdd4761bf1-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-kz9m9\" (UID: \"48b0efc4-5e8a-4bd9-9016-26bdd4761bf1\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-kz9m9" Dec 03 00:09:18 crc kubenswrapper[4953]: I1203 00:09:18.906242 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/73033ee8-cae5-4acc-8c77-60905eed62f2-mountpoint-dir\") pod \"csi-hostpathplugin-2xt2r\" (UID: \"73033ee8-cae5-4acc-8c77-60905eed62f2\") " pod="hostpath-provisioner/csi-hostpathplugin-2xt2r" Dec 03 00:09:18 crc kubenswrapper[4953]: I1203 00:09:18.906309 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nswff\" (UniqueName: \"kubernetes.io/projected/4edc57f1-5c0c-4108-bb3c-b2f104c6e08b-kube-api-access-nswff\") pod \"router-default-5444994796-r8l2l\" (UID: \"4edc57f1-5c0c-4108-bb3c-b2f104c6e08b\") " pod="openshift-ingress/router-default-5444994796-r8l2l" Dec 03 00:09:18 crc kubenswrapper[4953]: I1203 00:09:18.908037 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kcwc5\" (UniqueName: \"kubernetes.io/projected/536278ea-21b6-4193-ad9d-4dfb7ce6c194-kube-api-access-kcwc5\") pod \"package-server-manager-789f6589d5-dlqdp\" (UID: \"536278ea-21b6-4193-ad9d-4dfb7ce6c194\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-dlqdp" Dec 03 00:09:18 crc kubenswrapper[4953]: E1203 00:09:18.908618 4953 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 00:09:19.408599387 +0000 UTC m=+148.392880240 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-t7ljb" (UID: "4313395c-8501-4b5b-a4cb-2fc986fced03") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 00:09:18 crc kubenswrapper[4953]: I1203 00:09:18.909088 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/4313395c-8501-4b5b-a4cb-2fc986fced03-registry-certificates\") pod \"image-registry-697d97f7c8-t7ljb\" (UID: \"4313395c-8501-4b5b-a4cb-2fc986fced03\") " pod="openshift-image-registry/image-registry-697d97f7c8-t7ljb" Dec 03 00:09:18 crc kubenswrapper[4953]: I1203 00:09:18.921404 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6d44369d-4150-4832-9098-88d3c4b1133e-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-gdstj\" (UID: \"6d44369d-4150-4832-9098-88d3c4b1133e\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-gdstj" Dec 03 00:09:18 crc kubenswrapper[4953]: I1203 00:09:18.928775 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b77462cc-f0ef-4108-b81f-0d6d412d7fa8-serving-cert\") pod \"etcd-operator-b45778765-t968q\" (UID: \"b77462cc-f0ef-4108-b81f-0d6d412d7fa8\") " pod="openshift-etcd-operator/etcd-operator-b45778765-t968q" Dec 03 00:09:18 crc kubenswrapper[4953]: I1203 00:09:18.929539 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/90247294-af78-4d04-8aed-af90f93a034a-serving-cert\") pod \"route-controller-manager-6576b87f9c-52hrr\" (UID: \"90247294-af78-4d04-8aed-af90f93a034a\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-52hrr" Dec 03 00:09:18 crc kubenswrapper[4953]: I1203 00:09:18.929800 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-55l7h\" (UniqueName: \"kubernetes.io/projected/b77462cc-f0ef-4108-b81f-0d6d412d7fa8-kube-api-access-55l7h\") pod \"etcd-operator-b45778765-t968q\" (UID: \"b77462cc-f0ef-4108-b81f-0d6d412d7fa8\") " pod="openshift-etcd-operator/etcd-operator-b45778765-t968q" Dec 03 00:09:18 crc kubenswrapper[4953]: I1203 00:09:18.931192 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/a8a759e4-551c-41f9-bbab-9cfbf490777a-proxy-tls\") pod \"machine-config-controller-84d6567774-95vxq\" (UID: \"a8a759e4-551c-41f9-bbab-9cfbf490777a\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-95vxq" Dec 03 00:09:18 crc kubenswrapper[4953]: I1203 00:09:18.936050 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fwqw4\" (UniqueName: \"kubernetes.io/projected/90247294-af78-4d04-8aed-af90f93a034a-kube-api-access-fwqw4\") pod \"route-controller-manager-6576b87f9c-52hrr\" (UID: \"90247294-af78-4d04-8aed-af90f93a034a\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-52hrr" Dec 03 00:09:18 crc kubenswrapper[4953]: I1203 00:09:18.938301 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/2c280414-49b4-489f-b74a-13f6715744a6-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-l29cz\" (UID: \"2c280414-49b4-489f-b74a-13f6715744a6\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-l29cz" Dec 03 00:09:18 crc kubenswrapper[4953]: I1203 00:09:18.941574 4953 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-xbwlm"] Dec 03 00:09:18 crc kubenswrapper[4953]: I1203 00:09:18.946838 4953 patch_prober.go:28] interesting pod/machine-config-daemon-6dfbn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 00:09:18 crc kubenswrapper[4953]: I1203 00:09:18.946893 4953 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-6dfbn" podUID="53ffb1e9-9a2e-4869-80d7-776f4aa7ef8b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 00:09:18 crc kubenswrapper[4953]: I1203 00:09:18.958609 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c7jxx\" (UniqueName: \"kubernetes.io/projected/4313395c-8501-4b5b-a4cb-2fc986fced03-kube-api-access-c7jxx\") pod \"image-registry-697d97f7c8-t7ljb\" (UID: \"4313395c-8501-4b5b-a4cb-2fc986fced03\") " pod="openshift-image-registry/image-registry-697d97f7c8-t7ljb" Dec 03 00:09:18 crc kubenswrapper[4953]: I1203 00:09:18.963529 4953 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-45k44"] Dec 03 00:09:18 crc kubenswrapper[4953]: I1203 00:09:18.968320 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-dlqdp" Dec 03 00:09:18 crc kubenswrapper[4953]: I1203 00:09:18.975911 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dcpj6\" (UniqueName: \"kubernetes.io/projected/6d44369d-4150-4832-9098-88d3c4b1133e-kube-api-access-dcpj6\") pod \"openshift-apiserver-operator-796bbdcf4f-gdstj\" (UID: \"6d44369d-4150-4832-9098-88d3c4b1133e\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-gdstj" Dec 03 00:09:18 crc kubenswrapper[4953]: I1203 00:09:18.996947 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sbr4n\" (UniqueName: \"kubernetes.io/projected/2000e3b0-f8e7-4f0a-9f59-8c9d73324586-kube-api-access-sbr4n\") pod \"openshift-config-operator-7777fb866f-l7m9z\" (UID: \"2000e3b0-f8e7-4f0a-9f59-8c9d73324586\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-l7m9z" Dec 03 00:09:19 crc kubenswrapper[4953]: I1203 00:09:19.006864 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 00:09:19 crc kubenswrapper[4953]: I1203 00:09:19.007142 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/4edc57f1-5c0c-4108-bb3c-b2f104c6e08b-metrics-certs\") pod \"router-default-5444994796-r8l2l\" (UID: \"4edc57f1-5c0c-4108-bb3c-b2f104c6e08b\") " pod="openshift-ingress/router-default-5444994796-r8l2l" Dec 03 00:09:19 crc kubenswrapper[4953]: I1203 00:09:19.007186 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wgsd8\" (UniqueName: \"kubernetes.io/projected/4da5668c-a644-4014-8358-790da0f313e2-kube-api-access-wgsd8\") pod \"control-plane-machine-set-operator-78cbb6b69f-xhc8f\" (UID: \"4da5668c-a644-4014-8358-790da0f313e2\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-xhc8f" Dec 03 00:09:19 crc kubenswrapper[4953]: I1203 00:09:19.007214 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kkx26\" (UniqueName: \"kubernetes.io/projected/174354b2-d5cc-495a-ac13-cb56923aca08-kube-api-access-kkx26\") pod \"service-ca-9c57cc56f-zd9fs\" (UID: \"174354b2-d5cc-495a-ac13-cb56923aca08\") " pod="openshift-service-ca/service-ca-9c57cc56f-zd9fs" Dec 03 00:09:19 crc kubenswrapper[4953]: I1203 00:09:19.007236 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/1ee4772d-f4e3-4da8-a520-805dc2317fdd-profile-collector-cert\") pod \"catalog-operator-68c6474976-brcxn\" (UID: \"1ee4772d-f4e3-4da8-a520-805dc2317fdd\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-brcxn" Dec 03 00:09:19 crc kubenswrapper[4953]: I1203 00:09:19.007259 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/7124b81d-b87c-45de-a112-20602c4d9adb-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-kx9dt\" (UID: \"7124b81d-b87c-45de-a112-20602c4d9adb\") " pod="openshift-marketplace/marketplace-operator-79b997595-kx9dt" Dec 03 00:09:19 crc kubenswrapper[4953]: I1203 00:09:19.007285 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/48b0efc4-5e8a-4bd9-9016-26bdd4761bf1-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-kz9m9\" (UID: \"48b0efc4-5e8a-4bd9-9016-26bdd4761bf1\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-kz9m9" Dec 03 00:09:19 crc kubenswrapper[4953]: I1203 00:09:19.007309 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/13c8b8e5-a48a-4376-be38-f4902b55ad94-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-jfp8j\" (UID: \"13c8b8e5-a48a-4376-be38-f4902b55ad94\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-jfp8j" Dec 03 00:09:19 crc kubenswrapper[4953]: I1203 00:09:19.007328 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/73033ee8-cae5-4acc-8c77-60905eed62f2-csi-data-dir\") pod \"csi-hostpathplugin-2xt2r\" (UID: \"73033ee8-cae5-4acc-8c77-60905eed62f2\") " pod="hostpath-provisioner/csi-hostpathplugin-2xt2r" Dec 03 00:09:19 crc kubenswrapper[4953]: I1203 00:09:19.007353 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1f9a783f-9a9e-47b6-bc01-9c25261b1f2b-config-volume\") pod \"dns-default-qbsjg\" (UID: \"1f9a783f-9a9e-47b6-bc01-9c25261b1f2b\") " pod="openshift-dns/dns-default-qbsjg" Dec 03 00:09:19 crc kubenswrapper[4953]: I1203 00:09:19.007374 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1ffff94d-cd60-4079-86d6-ff10eb285ca2-config\") pod \"kube-apiserver-operator-766d6c64bb-z5glq\" (UID: \"1ffff94d-cd60-4079-86d6-ff10eb285ca2\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-z5glq" Dec 03 00:09:19 crc kubenswrapper[4953]: I1203 00:09:19.007394 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/c893caae-db46-4210-8be1-0ccf40bc9913-apiservice-cert\") pod \"packageserver-d55dfcdfc-qv29t\" (UID: \"c893caae-db46-4210-8be1-0ccf40bc9913\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-qv29t" Dec 03 00:09:19 crc kubenswrapper[4953]: I1203 00:09:19.007416 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5h6kc\" (UniqueName: \"kubernetes.io/projected/48b0efc4-5e8a-4bd9-9016-26bdd4761bf1-kube-api-access-5h6kc\") pod \"openshift-controller-manager-operator-756b6f6bc6-kz9m9\" (UID: \"48b0efc4-5e8a-4bd9-9016-26bdd4761bf1\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-kz9m9" Dec 03 00:09:19 crc kubenswrapper[4953]: I1203 00:09:19.007435 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/73033ee8-cae5-4acc-8c77-60905eed62f2-plugins-dir\") pod \"csi-hostpathplugin-2xt2r\" (UID: \"73033ee8-cae5-4acc-8c77-60905eed62f2\") " pod="hostpath-provisioner/csi-hostpathplugin-2xt2r" Dec 03 00:09:19 crc kubenswrapper[4953]: I1203 00:09:19.007453 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/4edc57f1-5c0c-4108-bb3c-b2f104c6e08b-default-certificate\") pod \"router-default-5444994796-r8l2l\" (UID: \"4edc57f1-5c0c-4108-bb3c-b2f104c6e08b\") " pod="openshift-ingress/router-default-5444994796-r8l2l" Dec 03 00:09:19 crc kubenswrapper[4953]: I1203 00:09:19.007473 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/c893caae-db46-4210-8be1-0ccf40bc9913-webhook-cert\") pod \"packageserver-d55dfcdfc-qv29t\" (UID: \"c893caae-db46-4210-8be1-0ccf40bc9913\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-qv29t" Dec 03 00:09:19 crc kubenswrapper[4953]: I1203 00:09:19.007494 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/e7e8210f-8915-49c3-8e95-f4427962f66a-srv-cert\") pod \"olm-operator-6b444d44fb-h95rc\" (UID: \"e7e8210f-8915-49c3-8e95-f4427962f66a\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-h95rc" Dec 03 00:09:19 crc kubenswrapper[4953]: I1203 00:09:19.007515 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/174354b2-d5cc-495a-ac13-cb56923aca08-signing-cabundle\") pod \"service-ca-9c57cc56f-zd9fs\" (UID: \"174354b2-d5cc-495a-ac13-cb56923aca08\") " pod="openshift-service-ca/service-ca-9c57cc56f-zd9fs" Dec 03 00:09:19 crc kubenswrapper[4953]: I1203 00:09:19.007537 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/fc4d1764-e8db-415b-9650-ed9a03cc49d1-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-njtzc\" (UID: \"fc4d1764-e8db-415b-9650-ed9a03cc49d1\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-njtzc" Dec 03 00:09:19 crc kubenswrapper[4953]: I1203 00:09:19.007556 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6whsj\" (UniqueName: \"kubernetes.io/projected/73033ee8-cae5-4acc-8c77-60905eed62f2-kube-api-access-6whsj\") pod \"csi-hostpathplugin-2xt2r\" (UID: \"73033ee8-cae5-4acc-8c77-60905eed62f2\") " pod="hostpath-provisioner/csi-hostpathplugin-2xt2r" Dec 03 00:09:19 crc kubenswrapper[4953]: I1203 00:09:19.007577 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/16c23c64-3f6c-4ce9-9118-cfeb890dfeed-config\") pod \"service-ca-operator-777779d784-fcw6n\" (UID: \"16c23c64-3f6c-4ce9-9118-cfeb890dfeed\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-fcw6n" Dec 03 00:09:19 crc kubenswrapper[4953]: I1203 00:09:19.007605 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7gsv5\" (UniqueName: \"kubernetes.io/projected/13c8b8e5-a48a-4376-be38-f4902b55ad94-kube-api-access-7gsv5\") pod \"kube-storage-version-migrator-operator-b67b599dd-jfp8j\" (UID: \"13c8b8e5-a48a-4376-be38-f4902b55ad94\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-jfp8j" Dec 03 00:09:19 crc kubenswrapper[4953]: I1203 00:09:19.007626 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1ffff94d-cd60-4079-86d6-ff10eb285ca2-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-z5glq\" (UID: \"1ffff94d-cd60-4079-86d6-ff10eb285ca2\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-z5glq" Dec 03 00:09:19 crc kubenswrapper[4953]: I1203 00:09:19.007647 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5bm5b\" (UniqueName: \"kubernetes.io/projected/7124b81d-b87c-45de-a112-20602c4d9adb-kube-api-access-5bm5b\") pod \"marketplace-operator-79b997595-kx9dt\" (UID: \"7124b81d-b87c-45de-a112-20602c4d9adb\") " pod="openshift-marketplace/marketplace-operator-79b997595-kx9dt" Dec 03 00:09:19 crc kubenswrapper[4953]: I1203 00:09:19.007679 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l7g4k\" (UniqueName: \"kubernetes.io/projected/1ee4772d-f4e3-4da8-a520-805dc2317fdd-kube-api-access-l7g4k\") pod \"catalog-operator-68c6474976-brcxn\" (UID: \"1ee4772d-f4e3-4da8-a520-805dc2317fdd\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-brcxn" Dec 03 00:09:19 crc kubenswrapper[4953]: I1203 00:09:19.007703 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/4edc57f1-5c0c-4108-bb3c-b2f104c6e08b-stats-auth\") pod \"router-default-5444994796-r8l2l\" (UID: \"4edc57f1-5c0c-4108-bb3c-b2f104c6e08b\") " pod="openshift-ingress/router-default-5444994796-r8l2l" Dec 03 00:09:19 crc kubenswrapper[4953]: I1203 00:09:19.007735 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/1f9a783f-9a9e-47b6-bc01-9c25261b1f2b-metrics-tls\") pod \"dns-default-qbsjg\" (UID: \"1f9a783f-9a9e-47b6-bc01-9c25261b1f2b\") " pod="openshift-dns/dns-default-qbsjg" Dec 03 00:09:19 crc kubenswrapper[4953]: I1203 00:09:19.007759 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-28pml\" (UniqueName: \"kubernetes.io/projected/16c23c64-3f6c-4ce9-9118-cfeb890dfeed-kube-api-access-28pml\") pod \"service-ca-operator-777779d784-fcw6n\" (UID: \"16c23c64-3f6c-4ce9-9118-cfeb890dfeed\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-fcw6n" Dec 03 00:09:19 crc kubenswrapper[4953]: I1203 00:09:19.007780 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/04f69b6e-9193-4055-8e50-e87fa663ca05-node-bootstrap-token\") pod \"machine-config-server-gpwzf\" (UID: \"04f69b6e-9193-4055-8e50-e87fa663ca05\") " pod="openshift-machine-config-operator/machine-config-server-gpwzf" Dec 03 00:09:19 crc kubenswrapper[4953]: I1203 00:09:19.007803 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/16c23c64-3f6c-4ce9-9118-cfeb890dfeed-serving-cert\") pod \"service-ca-operator-777779d784-fcw6n\" (UID: \"16c23c64-3f6c-4ce9-9118-cfeb890dfeed\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-fcw6n" Dec 03 00:09:19 crc kubenswrapper[4953]: I1203 00:09:19.007824 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/400790d0-9d4f-4698-bd68-8085ac29ec9c-cert\") pod \"ingress-canary-llvsw\" (UID: \"400790d0-9d4f-4698-bd68-8085ac29ec9c\") " pod="openshift-ingress-canary/ingress-canary-llvsw" Dec 03 00:09:19 crc kubenswrapper[4953]: I1203 00:09:19.007852 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/7b91526b-2492-4c17-949b-0ba3f8c221ea-config-volume\") pod \"collect-profiles-29412000-mqgxv\" (UID: \"7b91526b-2492-4c17-949b-0ba3f8c221ea\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412000-mqgxv" Dec 03 00:09:19 crc kubenswrapper[4953]: I1203 00:09:19.007877 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dvf62\" (UniqueName: \"kubernetes.io/projected/c893caae-db46-4210-8be1-0ccf40bc9913-kube-api-access-dvf62\") pod \"packageserver-d55dfcdfc-qv29t\" (UID: \"c893caae-db46-4210-8be1-0ccf40bc9913\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-qv29t" Dec 03 00:09:19 crc kubenswrapper[4953]: I1203 00:09:19.007897 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/7b91526b-2492-4c17-949b-0ba3f8c221ea-secret-volume\") pod \"collect-profiles-29412000-mqgxv\" (UID: \"7b91526b-2492-4c17-949b-0ba3f8c221ea\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412000-mqgxv" Dec 03 00:09:19 crc kubenswrapper[4953]: I1203 00:09:19.007940 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/1ee4772d-f4e3-4da8-a520-805dc2317fdd-srv-cert\") pod \"catalog-operator-68c6474976-brcxn\" (UID: \"1ee4772d-f4e3-4da8-a520-805dc2317fdd\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-brcxn" Dec 03 00:09:19 crc kubenswrapper[4953]: I1203 00:09:19.007965 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t4dws\" (UniqueName: \"kubernetes.io/projected/400790d0-9d4f-4698-bd68-8085ac29ec9c-kube-api-access-t4dws\") pod \"ingress-canary-llvsw\" (UID: \"400790d0-9d4f-4698-bd68-8085ac29ec9c\") " pod="openshift-ingress-canary/ingress-canary-llvsw" Dec 03 00:09:19 crc kubenswrapper[4953]: I1203 00:09:19.008010 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fc4d1764-e8db-415b-9650-ed9a03cc49d1-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-njtzc\" (UID: \"fc4d1764-e8db-415b-9650-ed9a03cc49d1\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-njtzc" Dec 03 00:09:19 crc kubenswrapper[4953]: I1203 00:09:19.008031 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/04f69b6e-9193-4055-8e50-e87fa663ca05-certs\") pod \"machine-config-server-gpwzf\" (UID: \"04f69b6e-9193-4055-8e50-e87fa663ca05\") " pod="openshift-machine-config-operator/machine-config-server-gpwzf" Dec 03 00:09:19 crc kubenswrapper[4953]: I1203 00:09:19.008070 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/c893caae-db46-4210-8be1-0ccf40bc9913-tmpfs\") pod \"packageserver-d55dfcdfc-qv29t\" (UID: \"c893caae-db46-4210-8be1-0ccf40bc9913\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-qv29t" Dec 03 00:09:19 crc kubenswrapper[4953]: I1203 00:09:19.008094 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/48b0efc4-5e8a-4bd9-9016-26bdd4761bf1-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-kz9m9\" (UID: \"48b0efc4-5e8a-4bd9-9016-26bdd4761bf1\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-kz9m9" Dec 03 00:09:19 crc kubenswrapper[4953]: I1203 00:09:19.008118 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/73033ee8-cae5-4acc-8c77-60905eed62f2-mountpoint-dir\") pod \"csi-hostpathplugin-2xt2r\" (UID: \"73033ee8-cae5-4acc-8c77-60905eed62f2\") " pod="hostpath-provisioner/csi-hostpathplugin-2xt2r" Dec 03 00:09:19 crc kubenswrapper[4953]: I1203 00:09:19.008142 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nswff\" (UniqueName: \"kubernetes.io/projected/4edc57f1-5c0c-4108-bb3c-b2f104c6e08b-kube-api-access-nswff\") pod \"router-default-5444994796-r8l2l\" (UID: \"4edc57f1-5c0c-4108-bb3c-b2f104c6e08b\") " pod="openshift-ingress/router-default-5444994796-r8l2l" Dec 03 00:09:19 crc kubenswrapper[4953]: I1203 00:09:19.008164 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/7124b81d-b87c-45de-a112-20602c4d9adb-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-kx9dt\" (UID: \"7124b81d-b87c-45de-a112-20602c4d9adb\") " pod="openshift-marketplace/marketplace-operator-79b997595-kx9dt" Dec 03 00:09:19 crc kubenswrapper[4953]: I1203 00:09:19.008187 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vzp6w\" (UniqueName: \"kubernetes.io/projected/04f69b6e-9193-4055-8e50-e87fa663ca05-kube-api-access-vzp6w\") pod \"machine-config-server-gpwzf\" (UID: \"04f69b6e-9193-4055-8e50-e87fa663ca05\") " pod="openshift-machine-config-operator/machine-config-server-gpwzf" Dec 03 00:09:19 crc kubenswrapper[4953]: I1203 00:09:19.008217 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6j5fc\" (UniqueName: \"kubernetes.io/projected/7b91526b-2492-4c17-949b-0ba3f8c221ea-kube-api-access-6j5fc\") pod \"collect-profiles-29412000-mqgxv\" (UID: \"7b91526b-2492-4c17-949b-0ba3f8c221ea\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412000-mqgxv" Dec 03 00:09:19 crc kubenswrapper[4953]: I1203 00:09:19.008245 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/73033ee8-cae5-4acc-8c77-60905eed62f2-socket-dir\") pod \"csi-hostpathplugin-2xt2r\" (UID: \"73033ee8-cae5-4acc-8c77-60905eed62f2\") " pod="hostpath-provisioner/csi-hostpathplugin-2xt2r" Dec 03 00:09:19 crc kubenswrapper[4953]: I1203 00:09:19.008267 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1ffff94d-cd60-4079-86d6-ff10eb285ca2-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-z5glq\" (UID: \"1ffff94d-cd60-4079-86d6-ff10eb285ca2\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-z5glq" Dec 03 00:09:19 crc kubenswrapper[4953]: I1203 00:09:19.008287 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/174354b2-d5cc-495a-ac13-cb56923aca08-signing-key\") pod \"service-ca-9c57cc56f-zd9fs\" (UID: \"174354b2-d5cc-495a-ac13-cb56923aca08\") " pod="openshift-service-ca/service-ca-9c57cc56f-zd9fs" Dec 03 00:09:19 crc kubenswrapper[4953]: I1203 00:09:19.008313 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7cs9l\" (UniqueName: \"kubernetes.io/projected/e7e8210f-8915-49c3-8e95-f4427962f66a-kube-api-access-7cs9l\") pod \"olm-operator-6b444d44fb-h95rc\" (UID: \"e7e8210f-8915-49c3-8e95-f4427962f66a\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-h95rc" Dec 03 00:09:19 crc kubenswrapper[4953]: I1203 00:09:19.008335 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xb6nh\" (UniqueName: \"kubernetes.io/projected/1f9a783f-9a9e-47b6-bc01-9c25261b1f2b-kube-api-access-xb6nh\") pod \"dns-default-qbsjg\" (UID: \"1f9a783f-9a9e-47b6-bc01-9c25261b1f2b\") " pod="openshift-dns/dns-default-qbsjg" Dec 03 00:09:19 crc kubenswrapper[4953]: I1203 00:09:19.008357 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/4edc57f1-5c0c-4108-bb3c-b2f104c6e08b-service-ca-bundle\") pod \"router-default-5444994796-r8l2l\" (UID: \"4edc57f1-5c0c-4108-bb3c-b2f104c6e08b\") " pod="openshift-ingress/router-default-5444994796-r8l2l" Dec 03 00:09:19 crc kubenswrapper[4953]: I1203 00:09:19.008409 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fc4d1764-e8db-415b-9650-ed9a03cc49d1-config\") pod \"kube-controller-manager-operator-78b949d7b-njtzc\" (UID: \"fc4d1764-e8db-415b-9650-ed9a03cc49d1\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-njtzc" Dec 03 00:09:19 crc kubenswrapper[4953]: I1203 00:09:19.008432 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/73033ee8-cae5-4acc-8c77-60905eed62f2-registration-dir\") pod \"csi-hostpathplugin-2xt2r\" (UID: \"73033ee8-cae5-4acc-8c77-60905eed62f2\") " pod="hostpath-provisioner/csi-hostpathplugin-2xt2r" Dec 03 00:09:19 crc kubenswrapper[4953]: I1203 00:09:19.008460 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/13c8b8e5-a48a-4376-be38-f4902b55ad94-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-jfp8j\" (UID: \"13c8b8e5-a48a-4376-be38-f4902b55ad94\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-jfp8j" Dec 03 00:09:19 crc kubenswrapper[4953]: I1203 00:09:19.008481 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sp7zr\" (UniqueName: \"kubernetes.io/projected/912368ff-455d-4a2d-a444-08c54448ca2a-kube-api-access-sp7zr\") pod \"migrator-59844c95c7-5fdxh\" (UID: \"912368ff-455d-4a2d-a444-08c54448ca2a\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-5fdxh" Dec 03 00:09:19 crc kubenswrapper[4953]: I1203 00:09:19.008505 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/e7e8210f-8915-49c3-8e95-f4427962f66a-profile-collector-cert\") pod \"olm-operator-6b444d44fb-h95rc\" (UID: \"e7e8210f-8915-49c3-8e95-f4427962f66a\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-h95rc" Dec 03 00:09:19 crc kubenswrapper[4953]: I1203 00:09:19.008528 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/4da5668c-a644-4014-8358-790da0f313e2-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-xhc8f\" (UID: \"4da5668c-a644-4014-8358-790da0f313e2\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-xhc8f" Dec 03 00:09:19 crc kubenswrapper[4953]: I1203 00:09:19.009149 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/73033ee8-cae5-4acc-8c77-60905eed62f2-plugins-dir\") pod \"csi-hostpathplugin-2xt2r\" (UID: \"73033ee8-cae5-4acc-8c77-60905eed62f2\") " pod="hostpath-provisioner/csi-hostpathplugin-2xt2r" Dec 03 00:09:19 crc kubenswrapper[4953]: I1203 00:09:19.009705 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/48b0efc4-5e8a-4bd9-9016-26bdd4761bf1-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-kz9m9\" (UID: \"48b0efc4-5e8a-4bd9-9016-26bdd4761bf1\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-kz9m9" Dec 03 00:09:19 crc kubenswrapper[4953]: E1203 00:09:19.009851 4953 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 00:09:19.509827397 +0000 UTC m=+148.494108180 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 00:09:19 crc kubenswrapper[4953]: I1203 00:09:19.010197 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/7124b81d-b87c-45de-a112-20602c4d9adb-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-kx9dt\" (UID: \"7124b81d-b87c-45de-a112-20602c4d9adb\") " pod="openshift-marketplace/marketplace-operator-79b997595-kx9dt" Dec 03 00:09:19 crc kubenswrapper[4953]: I1203 00:09:19.010594 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/73033ee8-cae5-4acc-8c77-60905eed62f2-socket-dir\") pod \"csi-hostpathplugin-2xt2r\" (UID: \"73033ee8-cae5-4acc-8c77-60905eed62f2\") " pod="hostpath-provisioner/csi-hostpathplugin-2xt2r" Dec 03 00:09:19 crc kubenswrapper[4953]: I1203 00:09:19.011678 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/16c23c64-3f6c-4ce9-9118-cfeb890dfeed-config\") pod \"service-ca-operator-777779d784-fcw6n\" (UID: \"16c23c64-3f6c-4ce9-9118-cfeb890dfeed\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-fcw6n" Dec 03 00:09:19 crc kubenswrapper[4953]: I1203 00:09:19.011776 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1ffff94d-cd60-4079-86d6-ff10eb285ca2-config\") pod \"kube-apiserver-operator-766d6c64bb-z5glq\" (UID: \"1ffff94d-cd60-4079-86d6-ff10eb285ca2\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-z5glq" Dec 03 00:09:19 crc kubenswrapper[4953]: I1203 00:09:19.011860 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/73033ee8-cae5-4acc-8c77-60905eed62f2-csi-data-dir\") pod \"csi-hostpathplugin-2xt2r\" (UID: \"73033ee8-cae5-4acc-8c77-60905eed62f2\") " pod="hostpath-provisioner/csi-hostpathplugin-2xt2r" Dec 03 00:09:19 crc kubenswrapper[4953]: I1203 00:09:19.011895 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/c893caae-db46-4210-8be1-0ccf40bc9913-tmpfs\") pod \"packageserver-d55dfcdfc-qv29t\" (UID: \"c893caae-db46-4210-8be1-0ccf40bc9913\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-qv29t" Dec 03 00:09:19 crc kubenswrapper[4953]: I1203 00:09:19.012955 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/13c8b8e5-a48a-4376-be38-f4902b55ad94-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-jfp8j\" (UID: \"13c8b8e5-a48a-4376-be38-f4902b55ad94\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-jfp8j" Dec 03 00:09:19 crc kubenswrapper[4953]: I1203 00:09:19.013118 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/4da5668c-a644-4014-8358-790da0f313e2-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-xhc8f\" (UID: \"4da5668c-a644-4014-8358-790da0f313e2\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-xhc8f" Dec 03 00:09:19 crc kubenswrapper[4953]: I1203 00:09:19.014347 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/4edc57f1-5c0c-4108-bb3c-b2f104c6e08b-service-ca-bundle\") pod \"router-default-5444994796-r8l2l\" (UID: \"4edc57f1-5c0c-4108-bb3c-b2f104c6e08b\") " pod="openshift-ingress/router-default-5444994796-r8l2l" Dec 03 00:09:19 crc kubenswrapper[4953]: I1203 00:09:19.014838 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/c893caae-db46-4210-8be1-0ccf40bc9913-webhook-cert\") pod \"packageserver-d55dfcdfc-qv29t\" (UID: \"c893caae-db46-4210-8be1-0ccf40bc9913\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-qv29t" Dec 03 00:09:19 crc kubenswrapper[4953]: I1203 00:09:19.015352 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fc4d1764-e8db-415b-9650-ed9a03cc49d1-config\") pod \"kube-controller-manager-operator-78b949d7b-njtzc\" (UID: \"fc4d1764-e8db-415b-9650-ed9a03cc49d1\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-njtzc" Dec 03 00:09:19 crc kubenswrapper[4953]: I1203 00:09:19.015461 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/73033ee8-cae5-4acc-8c77-60905eed62f2-registration-dir\") pod \"csi-hostpathplugin-2xt2r\" (UID: \"73033ee8-cae5-4acc-8c77-60905eed62f2\") " pod="hostpath-provisioner/csi-hostpathplugin-2xt2r" Dec 03 00:09:19 crc kubenswrapper[4953]: I1203 00:09:19.016700 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/c893caae-db46-4210-8be1-0ccf40bc9913-apiservice-cert\") pod \"packageserver-d55dfcdfc-qv29t\" (UID: \"c893caae-db46-4210-8be1-0ccf40bc9913\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-qv29t" Dec 03 00:09:19 crc kubenswrapper[4953]: I1203 00:09:19.017563 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/13c8b8e5-a48a-4376-be38-f4902b55ad94-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-jfp8j\" (UID: \"13c8b8e5-a48a-4376-be38-f4902b55ad94\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-jfp8j" Dec 03 00:09:19 crc kubenswrapper[4953]: I1203 00:09:19.018228 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/73033ee8-cae5-4acc-8c77-60905eed62f2-mountpoint-dir\") pod \"csi-hostpathplugin-2xt2r\" (UID: \"73033ee8-cae5-4acc-8c77-60905eed62f2\") " pod="hostpath-provisioner/csi-hostpathplugin-2xt2r" Dec 03 00:09:19 crc kubenswrapper[4953]: I1203 00:09:19.018811 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1f9a783f-9a9e-47b6-bc01-9c25261b1f2b-config-volume\") pod \"dns-default-qbsjg\" (UID: \"1f9a783f-9a9e-47b6-bc01-9c25261b1f2b\") " pod="openshift-dns/dns-default-qbsjg" Dec 03 00:09:19 crc kubenswrapper[4953]: I1203 00:09:19.019354 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/4edc57f1-5c0c-4108-bb3c-b2f104c6e08b-metrics-certs\") pod \"router-default-5444994796-r8l2l\" (UID: \"4edc57f1-5c0c-4108-bb3c-b2f104c6e08b\") " pod="openshift-ingress/router-default-5444994796-r8l2l" Dec 03 00:09:19 crc kubenswrapper[4953]: I1203 00:09:19.021307 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/7b91526b-2492-4c17-949b-0ba3f8c221ea-config-volume\") pod \"collect-profiles-29412000-mqgxv\" (UID: \"7b91526b-2492-4c17-949b-0ba3f8c221ea\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412000-mqgxv" Dec 03 00:09:19 crc kubenswrapper[4953]: I1203 00:09:19.021343 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/174354b2-d5cc-495a-ac13-cb56923aca08-signing-cabundle\") pod \"service-ca-9c57cc56f-zd9fs\" (UID: \"174354b2-d5cc-495a-ac13-cb56923aca08\") " pod="openshift-service-ca/service-ca-9c57cc56f-zd9fs" Dec 03 00:09:19 crc kubenswrapper[4953]: I1203 00:09:19.021835 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/48b0efc4-5e8a-4bd9-9016-26bdd4761bf1-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-kz9m9\" (UID: \"48b0efc4-5e8a-4bd9-9016-26bdd4761bf1\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-kz9m9" Dec 03 00:09:19 crc kubenswrapper[4953]: I1203 00:09:19.023522 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/04f69b6e-9193-4055-8e50-e87fa663ca05-node-bootstrap-token\") pod \"machine-config-server-gpwzf\" (UID: \"04f69b6e-9193-4055-8e50-e87fa663ca05\") " pod="openshift-machine-config-operator/machine-config-server-gpwzf" Dec 03 00:09:19 crc kubenswrapper[4953]: I1203 00:09:19.024159 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/1ee4772d-f4e3-4da8-a520-805dc2317fdd-profile-collector-cert\") pod \"catalog-operator-68c6474976-brcxn\" (UID: \"1ee4772d-f4e3-4da8-a520-805dc2317fdd\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-brcxn" Dec 03 00:09:19 crc kubenswrapper[4953]: I1203 00:09:19.026300 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/1f9a783f-9a9e-47b6-bc01-9c25261b1f2b-metrics-tls\") pod \"dns-default-qbsjg\" (UID: \"1f9a783f-9a9e-47b6-bc01-9c25261b1f2b\") " pod="openshift-dns/dns-default-qbsjg" Dec 03 00:09:19 crc kubenswrapper[4953]: I1203 00:09:19.028685 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/400790d0-9d4f-4698-bd68-8085ac29ec9c-cert\") pod \"ingress-canary-llvsw\" (UID: \"400790d0-9d4f-4698-bd68-8085ac29ec9c\") " pod="openshift-ingress-canary/ingress-canary-llvsw" Dec 03 00:09:19 crc kubenswrapper[4953]: I1203 00:09:19.029198 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/7124b81d-b87c-45de-a112-20602c4d9adb-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-kx9dt\" (UID: \"7124b81d-b87c-45de-a112-20602c4d9adb\") " pod="openshift-marketplace/marketplace-operator-79b997595-kx9dt" Dec 03 00:09:19 crc kubenswrapper[4953]: I1203 00:09:19.029127 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1ffff94d-cd60-4079-86d6-ff10eb285ca2-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-z5glq\" (UID: \"1ffff94d-cd60-4079-86d6-ff10eb285ca2\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-z5glq" Dec 03 00:09:19 crc kubenswrapper[4953]: I1203 00:09:19.039353 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/4edc57f1-5c0c-4108-bb3c-b2f104c6e08b-default-certificate\") pod \"router-default-5444994796-r8l2l\" (UID: \"4edc57f1-5c0c-4108-bb3c-b2f104c6e08b\") " pod="openshift-ingress/router-default-5444994796-r8l2l" Dec 03 00:09:19 crc kubenswrapper[4953]: I1203 00:09:19.039853 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/16c23c64-3f6c-4ce9-9118-cfeb890dfeed-serving-cert\") pod \"service-ca-operator-777779d784-fcw6n\" (UID: \"16c23c64-3f6c-4ce9-9118-cfeb890dfeed\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-fcw6n" Dec 03 00:09:19 crc kubenswrapper[4953]: I1203 00:09:19.040172 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/e7e8210f-8915-49c3-8e95-f4427962f66a-srv-cert\") pod \"olm-operator-6b444d44fb-h95rc\" (UID: \"e7e8210f-8915-49c3-8e95-f4427962f66a\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-h95rc" Dec 03 00:09:19 crc kubenswrapper[4953]: I1203 00:09:19.040270 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/secret/04f69b6e-9193-4055-8e50-e87fa663ca05-certs\") pod \"machine-config-server-gpwzf\" (UID: \"04f69b6e-9193-4055-8e50-e87fa663ca05\") " pod="openshift-machine-config-operator/machine-config-server-gpwzf" Dec 03 00:09:19 crc kubenswrapper[4953]: I1203 00:09:19.040297 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/1ee4772d-f4e3-4da8-a520-805dc2317fdd-srv-cert\") pod \"catalog-operator-68c6474976-brcxn\" (UID: \"1ee4772d-f4e3-4da8-a520-805dc2317fdd\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-brcxn" Dec 03 00:09:19 crc kubenswrapper[4953]: I1203 00:09:19.040434 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/e7e8210f-8915-49c3-8e95-f4427962f66a-profile-collector-cert\") pod \"olm-operator-6b444d44fb-h95rc\" (UID: \"e7e8210f-8915-49c3-8e95-f4427962f66a\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-h95rc" Dec 03 00:09:19 crc kubenswrapper[4953]: I1203 00:09:19.040942 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/174354b2-d5cc-495a-ac13-cb56923aca08-signing-key\") pod \"service-ca-9c57cc56f-zd9fs\" (UID: \"174354b2-d5cc-495a-ac13-cb56923aca08\") " pod="openshift-service-ca/service-ca-9c57cc56f-zd9fs" Dec 03 00:09:19 crc kubenswrapper[4953]: I1203 00:09:19.041280 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n5wnd\" (UniqueName: \"kubernetes.io/projected/4ce6dbe8-0292-4172-81a3-f5cdfdba6120-kube-api-access-n5wnd\") pod \"console-f9d7485db-m2msx\" (UID: \"4ce6dbe8-0292-4172-81a3-f5cdfdba6120\") " pod="openshift-console/console-f9d7485db-m2msx" Dec 03 00:09:19 crc kubenswrapper[4953]: I1203 00:09:19.042350 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fc4d1764-e8db-415b-9650-ed9a03cc49d1-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-njtzc\" (UID: \"fc4d1764-e8db-415b-9650-ed9a03cc49d1\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-njtzc" Dec 03 00:09:19 crc kubenswrapper[4953]: I1203 00:09:19.042873 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/7b91526b-2492-4c17-949b-0ba3f8c221ea-secret-volume\") pod \"collect-profiles-29412000-mqgxv\" (UID: \"7b91526b-2492-4c17-949b-0ba3f8c221ea\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412000-mqgxv" Dec 03 00:09:19 crc kubenswrapper[4953]: I1203 00:09:19.045813 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/4edc57f1-5c0c-4108-bb3c-b2f104c6e08b-stats-auth\") pod \"router-default-5444994796-r8l2l\" (UID: \"4edc57f1-5c0c-4108-bb3c-b2f104c6e08b\") " pod="openshift-ingress/router-default-5444994796-r8l2l" Dec 03 00:09:19 crc kubenswrapper[4953]: I1203 00:09:19.068917 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/4313395c-8501-4b5b-a4cb-2fc986fced03-bound-sa-token\") pod \"image-registry-697d97f7c8-t7ljb\" (UID: \"4313395c-8501-4b5b-a4cb-2fc986fced03\") " pod="openshift-image-registry/image-registry-697d97f7c8-t7ljb" Dec 03 00:09:19 crc kubenswrapper[4953]: I1203 00:09:19.070556 4953 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-c56d6"] Dec 03 00:09:19 crc kubenswrapper[4953]: W1203 00:09:19.071093 4953 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod496793fd_c6ab_48e4_a2c5_c259593c5d1c.slice/crio-bcbe6720c23ab2bf8172ced2359ced1cd48fc6e133a9f8bb509bb0adb73fa903 WatchSource:0}: Error finding container bcbe6720c23ab2bf8172ced2359ced1cd48fc6e133a9f8bb509bb0adb73fa903: Status 404 returned error can't find the container with id bcbe6720c23ab2bf8172ced2359ced1cd48fc6e133a9f8bb509bb0adb73fa903 Dec 03 00:09:19 crc kubenswrapper[4953]: I1203 00:09:19.072631 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h4grc\" (UniqueName: \"kubernetes.io/projected/a8a759e4-551c-41f9-bbab-9cfbf490777a-kube-api-access-h4grc\") pod \"machine-config-controller-84d6567774-95vxq\" (UID: \"a8a759e4-551c-41f9-bbab-9cfbf490777a\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-95vxq" Dec 03 00:09:19 crc kubenswrapper[4953]: I1203 00:09:19.076048 4953 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-hzbcl"] Dec 03 00:09:19 crc kubenswrapper[4953]: I1203 00:09:19.105658 4953 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-pruner-29412000-njkvg"] Dec 03 00:09:19 crc kubenswrapper[4953]: I1203 00:09:19.113282 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-t7ljb\" (UID: \"4313395c-8501-4b5b-a4cb-2fc986fced03\") " pod="openshift-image-registry/image-registry-697d97f7c8-t7ljb" Dec 03 00:09:19 crc kubenswrapper[4953]: E1203 00:09:19.113849 4953 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 00:09:19.613802955 +0000 UTC m=+148.598083918 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-t7ljb" (UID: "4313395c-8501-4b5b-a4cb-2fc986fced03") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 00:09:19 crc kubenswrapper[4953]: I1203 00:09:19.118829 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nswff\" (UniqueName: \"kubernetes.io/projected/4edc57f1-5c0c-4108-bb3c-b2f104c6e08b-kube-api-access-nswff\") pod \"router-default-5444994796-r8l2l\" (UID: \"4edc57f1-5c0c-4108-bb3c-b2f104c6e08b\") " pod="openshift-ingress/router-default-5444994796-r8l2l" Dec 03 00:09:19 crc kubenswrapper[4953]: I1203 00:09:19.134188 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-28pml\" (UniqueName: \"kubernetes.io/projected/16c23c64-3f6c-4ce9-9118-cfeb890dfeed-kube-api-access-28pml\") pod \"service-ca-operator-777779d784-fcw6n\" (UID: \"16c23c64-3f6c-4ce9-9118-cfeb890dfeed\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-fcw6n" Dec 03 00:09:19 crc kubenswrapper[4953]: I1203 00:09:19.163249 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7cs9l\" (UniqueName: \"kubernetes.io/projected/e7e8210f-8915-49c3-8e95-f4427962f66a-kube-api-access-7cs9l\") pod \"olm-operator-6b444d44fb-h95rc\" (UID: \"e7e8210f-8915-49c3-8e95-f4427962f66a\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-h95rc" Dec 03 00:09:19 crc kubenswrapper[4953]: I1203 00:09:19.167043 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-52hrr" Dec 03 00:09:19 crc kubenswrapper[4953]: I1203 00:09:19.174630 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vzp6w\" (UniqueName: \"kubernetes.io/projected/04f69b6e-9193-4055-8e50-e87fa663ca05-kube-api-access-vzp6w\") pod \"machine-config-server-gpwzf\" (UID: \"04f69b6e-9193-4055-8e50-e87fa663ca05\") " pod="openshift-machine-config-operator/machine-config-server-gpwzf" Dec 03 00:09:19 crc kubenswrapper[4953]: I1203 00:09:19.176170 4953 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-49rsf"] Dec 03 00:09:19 crc kubenswrapper[4953]: I1203 00:09:19.179648 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-2bh27" event={"ID":"8c657b07-c109-4475-bdd9-94808f3a58a0","Type":"ContainerStarted","Data":"f674b95f12080b55404905d893c332b7f340e4a929f680519238bcccd2a32299"} Dec 03 00:09:19 crc kubenswrapper[4953]: I1203 00:09:19.187041 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-45k44" event={"ID":"ca879198-0a45-4505-b845-60d8b0f5bb5d","Type":"ContainerStarted","Data":"7a2b12fa65694df3b4f7ebafb191325d197976df99a8b1dcf8e7a3a0a15276ed"} Dec 03 00:09:19 crc kubenswrapper[4953]: I1203 00:09:19.190677 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-l7sj9" event={"ID":"b33612b4-02d4-4f26-86bb-73a630d1ee58","Type":"ContainerStarted","Data":"2027e38371dd6192fed82e47860af430e69b879668794026dc7f3b6ba304f8e0"} Dec 03 00:09:19 crc kubenswrapper[4953]: I1203 00:09:19.190715 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-l7sj9" event={"ID":"b33612b4-02d4-4f26-86bb-73a630d1ee58","Type":"ContainerStarted","Data":"5a263737003b51158ff1c32ac37bbf8e444695110c8cf6d655006e58fc7c0938"} Dec 03 00:09:19 crc kubenswrapper[4953]: I1203 00:09:19.192109 4953 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-879f6c89f-l7sj9" Dec 03 00:09:19 crc kubenswrapper[4953]: I1203 00:09:19.192301 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-l29cz" Dec 03 00:09:19 crc kubenswrapper[4953]: I1203 00:09:19.193757 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6j5fc\" (UniqueName: \"kubernetes.io/projected/7b91526b-2492-4c17-949b-0ba3f8c221ea-kube-api-access-6j5fc\") pod \"collect-profiles-29412000-mqgxv\" (UID: \"7b91526b-2492-4c17-949b-0ba3f8c221ea\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412000-mqgxv" Dec 03 00:09:19 crc kubenswrapper[4953]: I1203 00:09:19.197227 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-5rzxr" event={"ID":"e0e3abc1-91d2-4a6a-aa1c-dc166f30d75a","Type":"ContainerStarted","Data":"266de1b2128802a0299a8dae46300c7d4c5411e9f59f0588048c25466c4df063"} Dec 03 00:09:19 crc kubenswrapper[4953]: I1203 00:09:19.201207 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-5rzxr" event={"ID":"e0e3abc1-91d2-4a6a-aa1c-dc166f30d75a","Type":"ContainerStarted","Data":"f77fc37ba4f9a607b3950e2f7018608a9c1cbce7dd845f6550acb1498eac8f2a"} Dec 03 00:09:19 crc kubenswrapper[4953]: I1203 00:09:19.201245 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-zmxjk" event={"ID":"f8f2cf57-42b1-4bb6-8c17-f9e0edb632fa","Type":"ContainerStarted","Data":"c347a1f63dbba8c8b0c61cd00b20034e8636f1bcac67b3df618d7c925f03ba61"} Dec 03 00:09:19 crc kubenswrapper[4953]: I1203 00:09:19.201265 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-zmxjk" event={"ID":"f8f2cf57-42b1-4bb6-8c17-f9e0edb632fa","Type":"ContainerStarted","Data":"70d2dfe4df0350fd764a0e2ccb891c0da3ca517987d5f34d3390e7b8389b2317"} Dec 03 00:09:19 crc kubenswrapper[4953]: I1203 00:09:19.201293 4953 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/downloads-7954f5f757-5rzxr" Dec 03 00:09:19 crc kubenswrapper[4953]: I1203 00:09:19.198558 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-m2msx" Dec 03 00:09:19 crc kubenswrapper[4953]: I1203 00:09:19.212545 4953 patch_prober.go:28] interesting pod/controller-manager-879f6c89f-l7sj9 container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.6:8443/healthz\": dial tcp 10.217.0.6:8443: connect: connection refused" start-of-body= Dec 03 00:09:19 crc kubenswrapper[4953]: I1203 00:09:19.212628 4953 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-879f6c89f-l7sj9" podUID="b33612b4-02d4-4f26-86bb-73a630d1ee58" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.6:8443/healthz\": dial tcp 10.217.0.6:8443: connect: connection refused" Dec 03 00:09:19 crc kubenswrapper[4953]: I1203 00:09:19.212897 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1ffff94d-cd60-4079-86d6-ff10eb285ca2-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-z5glq\" (UID: \"1ffff94d-cd60-4079-86d6-ff10eb285ca2\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-z5glq" Dec 03 00:09:19 crc kubenswrapper[4953]: I1203 00:09:19.216616 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-t968q" Dec 03 00:09:19 crc kubenswrapper[4953]: I1203 00:09:19.216869 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 00:09:19 crc kubenswrapper[4953]: E1203 00:09:19.217116 4953 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 00:09:19.717087356 +0000 UTC m=+148.701368139 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 00:09:19 crc kubenswrapper[4953]: I1203 00:09:19.217383 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-t7ljb\" (UID: \"4313395c-8501-4b5b-a4cb-2fc986fced03\") " pod="openshift-image-registry/image-registry-697d97f7c8-t7ljb" Dec 03 00:09:19 crc kubenswrapper[4953]: E1203 00:09:19.217745 4953 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 00:09:19.717730152 +0000 UTC m=+148.702010935 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-t7ljb" (UID: "4313395c-8501-4b5b-a4cb-2fc986fced03") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 00:09:19 crc kubenswrapper[4953]: I1203 00:09:19.218585 4953 patch_prober.go:28] interesting pod/downloads-7954f5f757-5rzxr container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" start-of-body= Dec 03 00:09:19 crc kubenswrapper[4953]: I1203 00:09:19.218624 4953 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-5rzxr" podUID="e0e3abc1-91d2-4a6a-aa1c-dc166f30d75a" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" Dec 03 00:09:19 crc kubenswrapper[4953]: I1203 00:09:19.219661 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-pruner-29412000-njkvg" event={"ID":"a68720c3-4f74-4435-91c6-99fd1b561416","Type":"ContainerStarted","Data":"a046b8835dc4b76faedddf1622c0469842261ac7a0761bfc7b1b596ad6b7814c"} Dec 03 00:09:19 crc kubenswrapper[4953]: I1203 00:09:19.223967 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-l7m9z" Dec 03 00:09:19 crc kubenswrapper[4953]: I1203 00:09:19.225086 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-xbwlm" event={"ID":"496793fd-c6ab-48e4-a2c5-c259593c5d1c","Type":"ContainerStarted","Data":"bcbe6720c23ab2bf8172ced2359ced1cd48fc6e133a9f8bb509bb0adb73fa903"} Dec 03 00:09:19 crc kubenswrapper[4953]: I1203 00:09:19.227273 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-c56d6" event={"ID":"8bd9553a-3081-43b8-a612-cee2a223cf77","Type":"ContainerStarted","Data":"10782125e7574eb624ae5d89de93404dac90d3503260c2462b09c28c7e9f0332"} Dec 03 00:09:19 crc kubenswrapper[4953]: I1203 00:09:19.235848 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xb6nh\" (UniqueName: \"kubernetes.io/projected/1f9a783f-9a9e-47b6-bc01-9c25261b1f2b-kube-api-access-xb6nh\") pod \"dns-default-qbsjg\" (UID: \"1f9a783f-9a9e-47b6-bc01-9c25261b1f2b\") " pod="openshift-dns/dns-default-qbsjg" Dec 03 00:09:19 crc kubenswrapper[4953]: I1203 00:09:19.253321 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wgsd8\" (UniqueName: \"kubernetes.io/projected/4da5668c-a644-4014-8358-790da0f313e2-kube-api-access-wgsd8\") pod \"control-plane-machine-set-operator-78cbb6b69f-xhc8f\" (UID: \"4da5668c-a644-4014-8358-790da0f313e2\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-xhc8f" Dec 03 00:09:19 crc kubenswrapper[4953]: I1203 00:09:19.255383 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-95vxq" Dec 03 00:09:19 crc kubenswrapper[4953]: I1203 00:09:19.262920 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-gdstj" Dec 03 00:09:19 crc kubenswrapper[4953]: I1203 00:09:19.273517 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kkx26\" (UniqueName: \"kubernetes.io/projected/174354b2-d5cc-495a-ac13-cb56923aca08-kube-api-access-kkx26\") pod \"service-ca-9c57cc56f-zd9fs\" (UID: \"174354b2-d5cc-495a-ac13-cb56923aca08\") " pod="openshift-service-ca/service-ca-9c57cc56f-zd9fs" Dec 03 00:09:19 crc kubenswrapper[4953]: I1203 00:09:19.286082 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-z5glq" Dec 03 00:09:19 crc kubenswrapper[4953]: I1203 00:09:19.292490 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-zd9fs" Dec 03 00:09:19 crc kubenswrapper[4953]: I1203 00:09:19.299611 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5h6kc\" (UniqueName: \"kubernetes.io/projected/48b0efc4-5e8a-4bd9-9016-26bdd4761bf1-kube-api-access-5h6kc\") pod \"openshift-controller-manager-operator-756b6f6bc6-kz9m9\" (UID: \"48b0efc4-5e8a-4bd9-9016-26bdd4761bf1\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-kz9m9" Dec 03 00:09:19 crc kubenswrapper[4953]: I1203 00:09:19.313556 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sp7zr\" (UniqueName: \"kubernetes.io/projected/912368ff-455d-4a2d-a444-08c54448ca2a-kube-api-access-sp7zr\") pod \"migrator-59844c95c7-5fdxh\" (UID: \"912368ff-455d-4a2d-a444-08c54448ca2a\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-5fdxh" Dec 03 00:09:19 crc kubenswrapper[4953]: I1203 00:09:19.320515 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 00:09:19 crc kubenswrapper[4953]: I1203 00:09:19.320827 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-5fdxh" Dec 03 00:09:19 crc kubenswrapper[4953]: E1203 00:09:19.321184 4953 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 00:09:19.821148916 +0000 UTC m=+148.805429699 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 00:09:19 crc kubenswrapper[4953]: I1203 00:09:19.321744 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-t7ljb\" (UID: \"4313395c-8501-4b5b-a4cb-2fc986fced03\") " pod="openshift-image-registry/image-registry-697d97f7c8-t7ljb" Dec 03 00:09:19 crc kubenswrapper[4953]: E1203 00:09:19.322422 4953 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 00:09:19.822391827 +0000 UTC m=+148.806672600 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-t7ljb" (UID: "4313395c-8501-4b5b-a4cb-2fc986fced03") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 00:09:19 crc kubenswrapper[4953]: I1203 00:09:19.326169 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-h95rc" Dec 03 00:09:19 crc kubenswrapper[4953]: I1203 00:09:19.334729 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dvf62\" (UniqueName: \"kubernetes.io/projected/c893caae-db46-4210-8be1-0ccf40bc9913-kube-api-access-dvf62\") pod \"packageserver-d55dfcdfc-qv29t\" (UID: \"c893caae-db46-4210-8be1-0ccf40bc9913\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-qv29t" Dec 03 00:09:19 crc kubenswrapper[4953]: I1203 00:09:19.340640 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-kz9m9" Dec 03 00:09:19 crc kubenswrapper[4953]: I1203 00:09:19.346554 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-r8l2l" Dec 03 00:09:19 crc kubenswrapper[4953]: I1203 00:09:19.352921 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t4dws\" (UniqueName: \"kubernetes.io/projected/400790d0-9d4f-4698-bd68-8085ac29ec9c-kube-api-access-t4dws\") pod \"ingress-canary-llvsw\" (UID: \"400790d0-9d4f-4698-bd68-8085ac29ec9c\") " pod="openshift-ingress-canary/ingress-canary-llvsw" Dec 03 00:09:19 crc kubenswrapper[4953]: I1203 00:09:19.355238 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-fcw6n" Dec 03 00:09:19 crc kubenswrapper[4953]: I1203 00:09:19.360637 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-xhc8f" Dec 03 00:09:19 crc kubenswrapper[4953]: I1203 00:09:19.367831 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-gpwzf" Dec 03 00:09:19 crc kubenswrapper[4953]: I1203 00:09:19.373571 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5bm5b\" (UniqueName: \"kubernetes.io/projected/7124b81d-b87c-45de-a112-20602c4d9adb-kube-api-access-5bm5b\") pod \"marketplace-operator-79b997595-kx9dt\" (UID: \"7124b81d-b87c-45de-a112-20602c4d9adb\") " pod="openshift-marketplace/marketplace-operator-79b997595-kx9dt" Dec 03 00:09:19 crc kubenswrapper[4953]: I1203 00:09:19.376905 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29412000-mqgxv" Dec 03 00:09:19 crc kubenswrapper[4953]: I1203 00:09:19.390945 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-qbsjg" Dec 03 00:09:19 crc kubenswrapper[4953]: I1203 00:09:19.401521 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7gsv5\" (UniqueName: \"kubernetes.io/projected/13c8b8e5-a48a-4376-be38-f4902b55ad94-kube-api-access-7gsv5\") pod \"kube-storage-version-migrator-operator-b67b599dd-jfp8j\" (UID: \"13c8b8e5-a48a-4376-be38-f4902b55ad94\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-jfp8j" Dec 03 00:09:19 crc kubenswrapper[4953]: I1203 00:09:19.405235 4953 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-vcx7x"] Dec 03 00:09:19 crc kubenswrapper[4953]: I1203 00:09:19.413884 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-llvsw" Dec 03 00:09:19 crc kubenswrapper[4953]: I1203 00:09:19.421079 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l7g4k\" (UniqueName: \"kubernetes.io/projected/1ee4772d-f4e3-4da8-a520-805dc2317fdd-kube-api-access-l7g4k\") pod \"catalog-operator-68c6474976-brcxn\" (UID: \"1ee4772d-f4e3-4da8-a520-805dc2317fdd\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-brcxn" Dec 03 00:09:19 crc kubenswrapper[4953]: I1203 00:09:19.423705 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 00:09:19 crc kubenswrapper[4953]: E1203 00:09:19.423864 4953 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 00:09:19.923828823 +0000 UTC m=+148.908109596 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 00:09:19 crc kubenswrapper[4953]: I1203 00:09:19.424021 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-t7ljb\" (UID: \"4313395c-8501-4b5b-a4cb-2fc986fced03\") " pod="openshift-image-registry/image-registry-697d97f7c8-t7ljb" Dec 03 00:09:19 crc kubenswrapper[4953]: E1203 00:09:19.424545 4953 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 00:09:19.92452973 +0000 UTC m=+148.908810523 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-t7ljb" (UID: "4313395c-8501-4b5b-a4cb-2fc986fced03") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 00:09:19 crc kubenswrapper[4953]: I1203 00:09:19.443677 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/fc4d1764-e8db-415b-9650-ed9a03cc49d1-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-njtzc\" (UID: \"fc4d1764-e8db-415b-9650-ed9a03cc49d1\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-njtzc" Dec 03 00:09:19 crc kubenswrapper[4953]: I1203 00:09:19.468765 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6whsj\" (UniqueName: \"kubernetes.io/projected/73033ee8-cae5-4acc-8c77-60905eed62f2-kube-api-access-6whsj\") pod \"csi-hostpathplugin-2xt2r\" (UID: \"73033ee8-cae5-4acc-8c77-60905eed62f2\") " pod="hostpath-provisioner/csi-hostpathplugin-2xt2r" Dec 03 00:09:19 crc kubenswrapper[4953]: I1203 00:09:19.538346 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 00:09:19 crc kubenswrapper[4953]: I1203 00:09:19.538522 4953 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-lhvnr"] Dec 03 00:09:19 crc kubenswrapper[4953]: E1203 00:09:19.538754 4953 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 00:09:20.038735638 +0000 UTC m=+149.023016431 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 00:09:19 crc kubenswrapper[4953]: I1203 00:09:19.544170 4953 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-gsszn"] Dec 03 00:09:19 crc kubenswrapper[4953]: I1203 00:09:19.555057 4953 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-lwq7b"] Dec 03 00:09:19 crc kubenswrapper[4953]: I1203 00:09:19.555121 4953 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-lg8jg"] Dec 03 00:09:19 crc kubenswrapper[4953]: I1203 00:09:19.563538 4953 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-8k8mf"] Dec 03 00:09:19 crc kubenswrapper[4953]: I1203 00:09:19.587359 4953 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-dlqdp"] Dec 03 00:09:19 crc kubenswrapper[4953]: I1203 00:09:19.599645 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-qv29t" Dec 03 00:09:19 crc kubenswrapper[4953]: I1203 00:09:19.609688 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-njtzc" Dec 03 00:09:19 crc kubenswrapper[4953]: I1203 00:09:19.625289 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-brcxn" Dec 03 00:09:19 crc kubenswrapper[4953]: I1203 00:09:19.634613 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-kx9dt" Dec 03 00:09:19 crc kubenswrapper[4953]: I1203 00:09:19.642443 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-t7ljb\" (UID: \"4313395c-8501-4b5b-a4cb-2fc986fced03\") " pod="openshift-image-registry/image-registry-697d97f7c8-t7ljb" Dec 03 00:09:19 crc kubenswrapper[4953]: E1203 00:09:19.642885 4953 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 00:09:20.14286991 +0000 UTC m=+149.127150693 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-t7ljb" (UID: "4313395c-8501-4b5b-a4cb-2fc986fced03") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 00:09:19 crc kubenswrapper[4953]: I1203 00:09:19.650320 4953 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-52hrr"] Dec 03 00:09:19 crc kubenswrapper[4953]: I1203 00:09:19.684583 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-jfp8j" Dec 03 00:09:19 crc kubenswrapper[4953]: I1203 00:09:19.703457 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-2xt2r" Dec 03 00:09:19 crc kubenswrapper[4953]: I1203 00:09:19.744484 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 00:09:19 crc kubenswrapper[4953]: E1203 00:09:19.744799 4953 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 00:09:20.244779398 +0000 UTC m=+149.229060181 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 00:09:19 crc kubenswrapper[4953]: W1203 00:09:19.820729 4953 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8a9efbc7_da01_4454_afce_64f004f240fd.slice/crio-65c0cb5ada8617c6fb47aa78dc4e5621c0f3c4fe4fc6d7eb02eefb520fddf45c WatchSource:0}: Error finding container 65c0cb5ada8617c6fb47aa78dc4e5621c0f3c4fe4fc6d7eb02eefb520fddf45c: Status 404 returned error can't find the container with id 65c0cb5ada8617c6fb47aa78dc4e5621c0f3c4fe4fc6d7eb02eefb520fddf45c Dec 03 00:09:19 crc kubenswrapper[4953]: W1203 00:09:19.831102 4953 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod536278ea_21b6_4193_ad9d_4dfb7ce6c194.slice/crio-fced7ae5a48d675bcd6f55817c7c9a661a52b485d93fe6db434d30980a47cb06 WatchSource:0}: Error finding container fced7ae5a48d675bcd6f55817c7c9a661a52b485d93fe6db434d30980a47cb06: Status 404 returned error can't find the container with id fced7ae5a48d675bcd6f55817c7c9a661a52b485d93fe6db434d30980a47cb06 Dec 03 00:09:19 crc kubenswrapper[4953]: I1203 00:09:19.848251 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-t7ljb\" (UID: \"4313395c-8501-4b5b-a4cb-2fc986fced03\") " pod="openshift-image-registry/image-registry-697d97f7c8-t7ljb" Dec 03 00:09:19 crc kubenswrapper[4953]: E1203 00:09:19.848709 4953 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 00:09:20.348692214 +0000 UTC m=+149.332972997 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-t7ljb" (UID: "4313395c-8501-4b5b-a4cb-2fc986fced03") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 00:09:19 crc kubenswrapper[4953]: W1203 00:09:19.870129 4953 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3b6479f0_333b_4a96_9adf_2099afdc2447.slice/crio-14485dd820a667c79bc3a331174667551289fbdb7cbaed9320c7e686363d59b4 WatchSource:0}: Error finding container 14485dd820a667c79bc3a331174667551289fbdb7cbaed9320c7e686363d59b4: Status 404 returned error can't find the container with id 14485dd820a667c79bc3a331174667551289fbdb7cbaed9320c7e686363d59b4 Dec 03 00:09:19 crc kubenswrapper[4953]: I1203 00:09:19.949848 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 00:09:19 crc kubenswrapper[4953]: E1203 00:09:19.954348 4953 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 00:09:20.454326413 +0000 UTC m=+149.438607196 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 00:09:20 crc kubenswrapper[4953]: I1203 00:09:20.044102 4953 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-95vxq"] Dec 03 00:09:20 crc kubenswrapper[4953]: I1203 00:09:20.054907 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-t7ljb\" (UID: \"4313395c-8501-4b5b-a4cb-2fc986fced03\") " pod="openshift-image-registry/image-registry-697d97f7c8-t7ljb" Dec 03 00:09:20 crc kubenswrapper[4953]: E1203 00:09:20.055350 4953 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 00:09:20.555333768 +0000 UTC m=+149.539614551 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-t7ljb" (UID: "4313395c-8501-4b5b-a4cb-2fc986fced03") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 00:09:20 crc kubenswrapper[4953]: I1203 00:09:20.162691 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 00:09:20 crc kubenswrapper[4953]: E1203 00:09:20.163397 4953 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 00:09:20.663370395 +0000 UTC m=+149.647651178 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 00:09:20 crc kubenswrapper[4953]: I1203 00:09:20.247425 4953 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-879f6c89f-l7sj9" podStartSLOduration=124.247396514 podStartE2EDuration="2m4.247396514s" podCreationTimestamp="2025-12-03 00:07:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 00:09:20.201010518 +0000 UTC m=+149.185291301" watchObservedRunningTime="2025-12-03 00:09:20.247396514 +0000 UTC m=+149.231677297" Dec 03 00:09:20 crc kubenswrapper[4953]: I1203 00:09:20.264707 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-t7ljb\" (UID: \"4313395c-8501-4b5b-a4cb-2fc986fced03\") " pod="openshift-image-registry/image-registry-697d97f7c8-t7ljb" Dec 03 00:09:20 crc kubenswrapper[4953]: E1203 00:09:20.265196 4953 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 00:09:20.7651734 +0000 UTC m=+149.749454183 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-t7ljb" (UID: "4313395c-8501-4b5b-a4cb-2fc986fced03") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 00:09:20 crc kubenswrapper[4953]: I1203 00:09:20.294937 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-52hrr" event={"ID":"90247294-af78-4d04-8aed-af90f93a034a","Type":"ContainerStarted","Data":"7526cbddebe178ec2171c80e10fe7a7efc3d2628ee342a668d58be317ca48031"} Dec 03 00:09:20 crc kubenswrapper[4953]: I1203 00:09:20.312668 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-hzbcl" event={"ID":"466d51cd-18a2-419a-a749-1ab91c943f13","Type":"ContainerStarted","Data":"ac8022bb8fa1ad1026b93d53f287f050979826086c64e63a79349b84028c865f"} Dec 03 00:09:20 crc kubenswrapper[4953]: I1203 00:09:20.317239 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"14485dd820a667c79bc3a331174667551289fbdb7cbaed9320c7e686363d59b4"} Dec 03 00:09:20 crc kubenswrapper[4953]: I1203 00:09:20.326331 4953 generic.go:334] "Generic (PLEG): container finished" podID="8c657b07-c109-4475-bdd9-94808f3a58a0" containerID="b83e2eb338ac5360c025c45a64495c677123c089338163908adb5554883b876e" exitCode=0 Dec 03 00:09:20 crc kubenswrapper[4953]: I1203 00:09:20.326403 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-2bh27" event={"ID":"8c657b07-c109-4475-bdd9-94808f3a58a0","Type":"ContainerDied","Data":"b83e2eb338ac5360c025c45a64495c677123c089338163908adb5554883b876e"} Dec 03 00:09:20 crc kubenswrapper[4953]: I1203 00:09:20.332709 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-r8l2l" event={"ID":"4edc57f1-5c0c-4108-bb3c-b2f104c6e08b","Type":"ContainerStarted","Data":"96051bb9c740f72b4d61804d5402649fc7230ff275f8df85d62729056e5f15dc"} Dec 03 00:09:20 crc kubenswrapper[4953]: I1203 00:09:20.334354 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-gsszn" event={"ID":"416a5985-9d2e-43f0-be89-d8be0ae77816","Type":"ContainerStarted","Data":"7a963d86ba50ad88cba14d5d70bb694ff9a5ca213ef5532feae823d53a020a8c"} Dec 03 00:09:20 crc kubenswrapper[4953]: I1203 00:09:20.335420 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-vcx7x" event={"ID":"6d99040d-450f-4ba8-a96d-16b694ca3636","Type":"ContainerStarted","Data":"1f00e003107744871318fde30008bc44c4b6ec23571c550029035cdce8d5f7e0"} Dec 03 00:09:20 crc kubenswrapper[4953]: I1203 00:09:20.336861 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-45k44" event={"ID":"ca879198-0a45-4505-b845-60d8b0f5bb5d","Type":"ContainerStarted","Data":"8da38403c4377107fa9a62218c1f47bc8b022b0acc0fc08c14bb8b6bffce93f6"} Dec 03 00:09:20 crc kubenswrapper[4953]: I1203 00:09:20.338203 4953 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-558db77b4-45k44" Dec 03 00:09:20 crc kubenswrapper[4953]: I1203 00:09:20.351506 4953 patch_prober.go:28] interesting pod/oauth-openshift-558db77b4-45k44 container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.8:6443/healthz\": dial tcp 10.217.0.8:6443: connect: connection refused" start-of-body= Dec 03 00:09:20 crc kubenswrapper[4953]: I1203 00:09:20.351568 4953 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-558db77b4-45k44" podUID="ca879198-0a45-4505-b845-60d8b0f5bb5d" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.8:6443/healthz\": dial tcp 10.217.0.8:6443: connect: connection refused" Dec 03 00:09:20 crc kubenswrapper[4953]: I1203 00:09:20.353115 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-dlqdp" event={"ID":"536278ea-21b6-4193-ad9d-4dfb7ce6c194","Type":"ContainerStarted","Data":"fced7ae5a48d675bcd6f55817c7c9a661a52b485d93fe6db434d30980a47cb06"} Dec 03 00:09:20 crc kubenswrapper[4953]: I1203 00:09:20.367474 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 00:09:20 crc kubenswrapper[4953]: E1203 00:09:20.369052 4953 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 00:09:20.869014225 +0000 UTC m=+149.853295008 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 00:09:20 crc kubenswrapper[4953]: I1203 00:09:20.378004 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-lhvnr" event={"ID":"98788d99-bf86-41b8-b71b-0c1edab5dca2","Type":"ContainerStarted","Data":"c3cf38e5d65dfc5b827f53f49f2ff34178de02d5532df2d97881b389b2efd233"} Dec 03 00:09:20 crc kubenswrapper[4953]: I1203 00:09:20.381671 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-t7ljb\" (UID: \"4313395c-8501-4b5b-a4cb-2fc986fced03\") " pod="openshift-image-registry/image-registry-697d97f7c8-t7ljb" Dec 03 00:09:20 crc kubenswrapper[4953]: E1203 00:09:20.382556 4953 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 00:09:20.882531056 +0000 UTC m=+149.866811839 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-t7ljb" (UID: "4313395c-8501-4b5b-a4cb-2fc986fced03") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 00:09:20 crc kubenswrapper[4953]: I1203 00:09:20.430060 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-xbwlm" event={"ID":"496793fd-c6ab-48e4-a2c5-c259593c5d1c","Type":"ContainerStarted","Data":"3c4923c1e987bbd06f6ea091e939c6d4ee307a2479b2d237b6e0f43f6769cc10"} Dec 03 00:09:20 crc kubenswrapper[4953]: I1203 00:09:20.462921 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-c56d6" event={"ID":"8bd9553a-3081-43b8-a612-cee2a223cf77","Type":"ContainerStarted","Data":"d47307aa91d7235c92db393b08b589a3c488e18c4c91211511be427863636769"} Dec 03 00:09:20 crc kubenswrapper[4953]: I1203 00:09:20.463712 4953 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console-operator/console-operator-58897d9998-c56d6" Dec 03 00:09:20 crc kubenswrapper[4953]: I1203 00:09:20.468925 4953 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/downloads-7954f5f757-5rzxr" podStartSLOduration=124.468890292 podStartE2EDuration="2m4.468890292s" podCreationTimestamp="2025-12-03 00:07:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 00:09:20.468115143 +0000 UTC m=+149.452395926" watchObservedRunningTime="2025-12-03 00:09:20.468890292 +0000 UTC m=+149.453171075" Dec 03 00:09:20 crc kubenswrapper[4953]: I1203 00:09:20.477295 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"f9d5153089fd9a98048aac6fc8b46700c4503260321268b0f1c3d9591f6b6cb6"} Dec 03 00:09:20 crc kubenswrapper[4953]: I1203 00:09:20.478837 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-gpwzf" event={"ID":"04f69b6e-9193-4055-8e50-e87fa663ca05","Type":"ContainerStarted","Data":"347dec1fe6ef98230b0fe0106352cd20e676aca77f721742e3fbcb27ec2515f5"} Dec 03 00:09:20 crc kubenswrapper[4953]: I1203 00:09:20.479489 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-lg8jg" event={"ID":"8a9efbc7-da01-4454-afce-64f004f240fd","Type":"ContainerStarted","Data":"65c0cb5ada8617c6fb47aa78dc4e5621c0f3c4fe4fc6d7eb02eefb520fddf45c"} Dec 03 00:09:20 crc kubenswrapper[4953]: I1203 00:09:20.480153 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-lwq7b" event={"ID":"a7477307-0aa0-461d-8fc7-f4f8fc2f7c7b","Type":"ContainerStarted","Data":"a26087af688dd0ad6b54de7b63c1f61948babae0be9e22849eb86f720502e364"} Dec 03 00:09:20 crc kubenswrapper[4953]: I1203 00:09:20.482503 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 00:09:20 crc kubenswrapper[4953]: E1203 00:09:20.483409 4953 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 00:09:20.983388437 +0000 UTC m=+149.967669230 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 00:09:20 crc kubenswrapper[4953]: I1203 00:09:20.487820 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-8k8mf" event={"ID":"b2f23cca-bfc0-4e7a-bf15-2fc4065e840b","Type":"ContainerStarted","Data":"54d52419d3ee3ac6d0880a436e95a2f4e6dd8abcb0a4371e2edf09821e1944de"} Dec 03 00:09:20 crc kubenswrapper[4953]: I1203 00:09:20.503024 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-pruner-29412000-njkvg" event={"ID":"a68720c3-4f74-4435-91c6-99fd1b561416","Type":"ContainerStarted","Data":"ce83851e48b19dc51f2a638a8c17cff4505f4f43bd73cda58510bf0e9d803ffa"} Dec 03 00:09:20 crc kubenswrapper[4953]: I1203 00:09:20.514356 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-zmxjk" event={"ID":"f8f2cf57-42b1-4bb6-8c17-f9e0edb632fa","Type":"ContainerStarted","Data":"5ca201c9275a76c96c831a1d959cc4bc770d70760574cc617f7c6b5b2b5d2992"} Dec 03 00:09:20 crc kubenswrapper[4953]: I1203 00:09:20.516368 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-49rsf" event={"ID":"0d6c7577-4632-4e4d-88f0-87548a6c640f","Type":"ContainerStarted","Data":"b1340ee1c1ed8bbd46ae4de2623c56c37909f7ff7e7965aee57ea7f2fb686a10"} Dec 03 00:09:20 crc kubenswrapper[4953]: I1203 00:09:20.517277 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"c74ef136e8d740efc4446703e0f91620a513889b10f2fa1c4cea10e02e7ab66c"} Dec 03 00:09:20 crc kubenswrapper[4953]: I1203 00:09:20.517902 4953 patch_prober.go:28] interesting pod/downloads-7954f5f757-5rzxr container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" start-of-body= Dec 03 00:09:20 crc kubenswrapper[4953]: I1203 00:09:20.517933 4953 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-5rzxr" podUID="e0e3abc1-91d2-4a6a-aa1c-dc166f30d75a" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" Dec 03 00:09:20 crc kubenswrapper[4953]: I1203 00:09:20.591962 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-t7ljb\" (UID: \"4313395c-8501-4b5b-a4cb-2fc986fced03\") " pod="openshift-image-registry/image-registry-697d97f7c8-t7ljb" Dec 03 00:09:20 crc kubenswrapper[4953]: E1203 00:09:20.592319 4953 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 00:09:21.092304286 +0000 UTC m=+150.076585069 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-t7ljb" (UID: "4313395c-8501-4b5b-a4cb-2fc986fced03") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 00:09:20 crc kubenswrapper[4953]: I1203 00:09:20.693692 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 00:09:20 crc kubenswrapper[4953]: E1203 00:09:20.693958 4953 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 00:09:21.193916476 +0000 UTC m=+150.178197259 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 00:09:20 crc kubenswrapper[4953]: I1203 00:09:20.694136 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-t7ljb\" (UID: \"4313395c-8501-4b5b-a4cb-2fc986fced03\") " pod="openshift-image-registry/image-registry-697d97f7c8-t7ljb" Dec 03 00:09:20 crc kubenswrapper[4953]: E1203 00:09:20.694407 4953 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 00:09:21.194399828 +0000 UTC m=+150.178680611 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-t7ljb" (UID: "4313395c-8501-4b5b-a4cb-2fc986fced03") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 00:09:20 crc kubenswrapper[4953]: I1203 00:09:20.733158 4953 patch_prober.go:28] interesting pod/console-operator-58897d9998-c56d6 container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.7:8443/readyz\": dial tcp 10.217.0.7:8443: connect: connection refused" start-of-body= Dec 03 00:09:20 crc kubenswrapper[4953]: I1203 00:09:20.733227 4953 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-c56d6" podUID="8bd9553a-3081-43b8-a612-cee2a223cf77" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.7:8443/readyz\": dial tcp 10.217.0.7:8443: connect: connection refused" Dec 03 00:09:20 crc kubenswrapper[4953]: I1203 00:09:20.745257 4953 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-879f6c89f-l7sj9" Dec 03 00:09:20 crc kubenswrapper[4953]: I1203 00:09:20.873057 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 00:09:20 crc kubenswrapper[4953]: E1203 00:09:20.873520 4953 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 00:09:21.373498837 +0000 UTC m=+150.357779620 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 00:09:20 crc kubenswrapper[4953]: I1203 00:09:20.873806 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-t7ljb\" (UID: \"4313395c-8501-4b5b-a4cb-2fc986fced03\") " pod="openshift-image-registry/image-registry-697d97f7c8-t7ljb" Dec 03 00:09:20 crc kubenswrapper[4953]: E1203 00:09:20.874218 4953 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 00:09:21.374208244 +0000 UTC m=+150.358489027 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-t7ljb" (UID: "4313395c-8501-4b5b-a4cb-2fc986fced03") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 00:09:20 crc kubenswrapper[4953]: I1203 00:09:20.974960 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 00:09:20 crc kubenswrapper[4953]: E1203 00:09:20.975278 4953 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 00:09:21.475261691 +0000 UTC m=+150.459542474 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 00:09:21 crc kubenswrapper[4953]: I1203 00:09:21.083653 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-t7ljb\" (UID: \"4313395c-8501-4b5b-a4cb-2fc986fced03\") " pod="openshift-image-registry/image-registry-697d97f7c8-t7ljb" Dec 03 00:09:21 crc kubenswrapper[4953]: E1203 00:09:21.084901 4953 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 00:09:21.584885577 +0000 UTC m=+150.569166360 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-t7ljb" (UID: "4313395c-8501-4b5b-a4cb-2fc986fced03") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 00:09:21 crc kubenswrapper[4953]: I1203 00:09:21.190430 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 00:09:21 crc kubenswrapper[4953]: E1203 00:09:21.190606 4953 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 00:09:21.690576587 +0000 UTC m=+150.674857370 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 00:09:21 crc kubenswrapper[4953]: I1203 00:09:21.190682 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-t7ljb\" (UID: \"4313395c-8501-4b5b-a4cb-2fc986fced03\") " pod="openshift-image-registry/image-registry-697d97f7c8-t7ljb" Dec 03 00:09:21 crc kubenswrapper[4953]: E1203 00:09:21.191006 4953 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 00:09:21.690995127 +0000 UTC m=+150.675275900 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-t7ljb" (UID: "4313395c-8501-4b5b-a4cb-2fc986fced03") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 00:09:21 crc kubenswrapper[4953]: I1203 00:09:21.294051 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 00:09:21 crc kubenswrapper[4953]: E1203 00:09:21.294484 4953 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 00:09:21.794461173 +0000 UTC m=+150.778741956 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 00:09:21 crc kubenswrapper[4953]: I1203 00:09:21.395698 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-t7ljb\" (UID: \"4313395c-8501-4b5b-a4cb-2fc986fced03\") " pod="openshift-image-registry/image-registry-697d97f7c8-t7ljb" Dec 03 00:09:21 crc kubenswrapper[4953]: E1203 00:09:21.396014 4953 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 00:09:21.896000991 +0000 UTC m=+150.880281774 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-t7ljb" (UID: "4313395c-8501-4b5b-a4cb-2fc986fced03") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 00:09:21 crc kubenswrapper[4953]: I1203 00:09:21.518473 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 00:09:21 crc kubenswrapper[4953]: E1203 00:09:21.519019 4953 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 00:09:22.019002775 +0000 UTC m=+151.003283558 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 00:09:21 crc kubenswrapper[4953]: I1203 00:09:21.543672 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"208e2c9a8c05c0dfd510a342dca1f1c511ad266e8fa85d6f4cae8f51c5634942"} Dec 03 00:09:21 crc kubenswrapper[4953]: I1203 00:09:21.550555 4953 generic.go:334] "Generic (PLEG): container finished" podID="466d51cd-18a2-419a-a749-1ab91c943f13" containerID="acab4105bd0b6f355d984b0d95ce2bd1c8e27f213bd9106d3115fc3571c83f36" exitCode=0 Dec 03 00:09:21 crc kubenswrapper[4953]: I1203 00:09:21.551366 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-hzbcl" event={"ID":"466d51cd-18a2-419a-a749-1ab91c943f13","Type":"ContainerDied","Data":"acab4105bd0b6f355d984b0d95ce2bd1c8e27f213bd9106d3115fc3571c83f36"} Dec 03 00:09:21 crc kubenswrapper[4953]: I1203 00:09:21.564874 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-95vxq" event={"ID":"a8a759e4-551c-41f9-bbab-9cfbf490777a","Type":"ContainerStarted","Data":"b59ae1c0001757638dd5e43d2c47a023c5cbca48a6d722de9a8d618793f6d3d7"} Dec 03 00:09:21 crc kubenswrapper[4953]: I1203 00:09:21.568020 4953 patch_prober.go:28] interesting pod/console-operator-58897d9998-c56d6 container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.7:8443/readyz\": dial tcp 10.217.0.7:8443: connect: connection refused" start-of-body= Dec 03 00:09:21 crc kubenswrapper[4953]: I1203 00:09:21.568077 4953 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-c56d6" podUID="8bd9553a-3081-43b8-a612-cee2a223cf77" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.7:8443/readyz\": dial tcp 10.217.0.7:8443: connect: connection refused" Dec 03 00:09:21 crc kubenswrapper[4953]: I1203 00:09:21.611630 4953 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-558db77b4-45k44" Dec 03 00:09:21 crc kubenswrapper[4953]: I1203 00:09:21.619839 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-t7ljb\" (UID: \"4313395c-8501-4b5b-a4cb-2fc986fced03\") " pod="openshift-image-registry/image-registry-697d97f7c8-t7ljb" Dec 03 00:09:21 crc kubenswrapper[4953]: E1203 00:09:21.620205 4953 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 00:09:22.120193995 +0000 UTC m=+151.104474778 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-t7ljb" (UID: "4313395c-8501-4b5b-a4cb-2fc986fced03") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 00:09:21 crc kubenswrapper[4953]: I1203 00:09:21.720875 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 00:09:21 crc kubenswrapper[4953]: E1203 00:09:21.722056 4953 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 00:09:22.222040891 +0000 UTC m=+151.206321674 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 00:09:21 crc kubenswrapper[4953]: I1203 00:09:21.737955 4953 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-pruner-29412000-njkvg" podStartSLOduration=126.73792684 podStartE2EDuration="2m6.73792684s" podCreationTimestamp="2025-12-03 00:07:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 00:09:21.65876636 +0000 UTC m=+150.643047133" watchObservedRunningTime="2025-12-03 00:09:21.73792684 +0000 UTC m=+150.722207623" Dec 03 00:09:21 crc kubenswrapper[4953]: I1203 00:09:21.745660 4953 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-zmxjk" podStartSLOduration=126.745641739 podStartE2EDuration="2m6.745641739s" podCreationTimestamp="2025-12-03 00:07:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 00:09:21.742676366 +0000 UTC m=+150.726957149" watchObservedRunningTime="2025-12-03 00:09:21.745641739 +0000 UTC m=+150.729922522" Dec 03 00:09:21 crc kubenswrapper[4953]: I1203 00:09:21.822797 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-t7ljb\" (UID: \"4313395c-8501-4b5b-a4cb-2fc986fced03\") " pod="openshift-image-registry/image-registry-697d97f7c8-t7ljb" Dec 03 00:09:21 crc kubenswrapper[4953]: E1203 00:09:21.823101 4953 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 00:09:22.323090197 +0000 UTC m=+151.307370980 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-t7ljb" (UID: "4313395c-8501-4b5b-a4cb-2fc986fced03") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 00:09:21 crc kubenswrapper[4953]: I1203 00:09:21.924282 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 00:09:21 crc kubenswrapper[4953]: E1203 00:09:21.924504 4953 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 00:09:22.424464141 +0000 UTC m=+151.408744954 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 00:09:21 crc kubenswrapper[4953]: I1203 00:09:21.924690 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-t7ljb\" (UID: \"4313395c-8501-4b5b-a4cb-2fc986fced03\") " pod="openshift-image-registry/image-registry-697d97f7c8-t7ljb" Dec 03 00:09:21 crc kubenswrapper[4953]: E1203 00:09:21.925193 4953 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 00:09:22.425170048 +0000 UTC m=+151.409450831 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-t7ljb" (UID: "4313395c-8501-4b5b-a4cb-2fc986fced03") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 00:09:21 crc kubenswrapper[4953]: I1203 00:09:21.933936 4953 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console-operator/console-operator-58897d9998-c56d6" podStartSLOduration=125.933915622 podStartE2EDuration="2m5.933915622s" podCreationTimestamp="2025-12-03 00:07:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 00:09:21.781000855 +0000 UTC m=+150.765281648" watchObservedRunningTime="2025-12-03 00:09:21.933915622 +0000 UTC m=+150.918196405" Dec 03 00:09:21 crc kubenswrapper[4953]: I1203 00:09:21.974650 4953 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-558db77b4-45k44" podStartSLOduration=126.97462832 podStartE2EDuration="2m6.97462832s" podCreationTimestamp="2025-12-03 00:07:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 00:09:21.972201171 +0000 UTC m=+150.956481954" watchObservedRunningTime="2025-12-03 00:09:21.97462832 +0000 UTC m=+150.958909103" Dec 03 00:09:22 crc kubenswrapper[4953]: I1203 00:09:22.026067 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 00:09:22 crc kubenswrapper[4953]: E1203 00:09:22.026344 4953 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 00:09:22.526294686 +0000 UTC m=+151.510575469 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 00:09:22 crc kubenswrapper[4953]: I1203 00:09:22.026568 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-t7ljb\" (UID: \"4313395c-8501-4b5b-a4cb-2fc986fced03\") " pod="openshift-image-registry/image-registry-697d97f7c8-t7ljb" Dec 03 00:09:22 crc kubenswrapper[4953]: E1203 00:09:22.027050 4953 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 00:09:22.527042075 +0000 UTC m=+151.511322858 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-t7ljb" (UID: "4313395c-8501-4b5b-a4cb-2fc986fced03") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 00:09:22 crc kubenswrapper[4953]: I1203 00:09:22.132790 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 00:09:22 crc kubenswrapper[4953]: E1203 00:09:22.141125 4953 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 00:09:22.641096349 +0000 UTC m=+151.625377132 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 00:09:22 crc kubenswrapper[4953]: I1203 00:09:22.243814 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-t7ljb\" (UID: \"4313395c-8501-4b5b-a4cb-2fc986fced03\") " pod="openshift-image-registry/image-registry-697d97f7c8-t7ljb" Dec 03 00:09:22 crc kubenswrapper[4953]: E1203 00:09:22.244242 4953 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 00:09:22.744225037 +0000 UTC m=+151.728505820 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-t7ljb" (UID: "4313395c-8501-4b5b-a4cb-2fc986fced03") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 00:09:22 crc kubenswrapper[4953]: I1203 00:09:22.345012 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 00:09:22 crc kubenswrapper[4953]: E1203 00:09:22.345382 4953 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 00:09:22.845360214 +0000 UTC m=+151.829640997 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 00:09:22 crc kubenswrapper[4953]: I1203 00:09:22.345464 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-t7ljb\" (UID: \"4313395c-8501-4b5b-a4cb-2fc986fced03\") " pod="openshift-image-registry/image-registry-697d97f7c8-t7ljb" Dec 03 00:09:22 crc kubenswrapper[4953]: E1203 00:09:22.345783 4953 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 00:09:22.845768944 +0000 UTC m=+151.830049727 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-t7ljb" (UID: "4313395c-8501-4b5b-a4cb-2fc986fced03") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 00:09:22 crc kubenswrapper[4953]: I1203 00:09:22.350245 4953 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-l29cz"] Dec 03 00:09:22 crc kubenswrapper[4953]: I1203 00:09:22.450255 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 00:09:22 crc kubenswrapper[4953]: E1203 00:09:22.450578 4953 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 00:09:22.950560162 +0000 UTC m=+151.934840945 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 00:09:22 crc kubenswrapper[4953]: I1203 00:09:22.554360 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-t7ljb\" (UID: \"4313395c-8501-4b5b-a4cb-2fc986fced03\") " pod="openshift-image-registry/image-registry-697d97f7c8-t7ljb" Dec 03 00:09:22 crc kubenswrapper[4953]: E1203 00:09:22.555403 4953 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 00:09:23.055386191 +0000 UTC m=+152.039666974 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-t7ljb" (UID: "4313395c-8501-4b5b-a4cb-2fc986fced03") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 00:09:22 crc kubenswrapper[4953]: I1203 00:09:22.656469 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 00:09:22 crc kubenswrapper[4953]: E1203 00:09:22.656821 4953 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 00:09:23.156785366 +0000 UTC m=+152.141066149 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 00:09:22 crc kubenswrapper[4953]: I1203 00:09:22.687827 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-r8l2l" event={"ID":"4edc57f1-5c0c-4108-bb3c-b2f104c6e08b","Type":"ContainerStarted","Data":"91be52237eab3cd63e1073cb8af53d8266f01253168fbb3281bd056f3be44c7d"} Dec 03 00:09:22 crc kubenswrapper[4953]: I1203 00:09:22.689829 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"6ef3ef369c5cf340d658c19a303858d75367d83651c39f987b850f0dfd3d8c50"} Dec 03 00:09:22 crc kubenswrapper[4953]: I1203 00:09:22.690501 4953 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 00:09:22 crc kubenswrapper[4953]: I1203 00:09:22.691857 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-95vxq" event={"ID":"a8a759e4-551c-41f9-bbab-9cfbf490777a","Type":"ContainerStarted","Data":"25206c33d07a87b59a7392db16dda3336dacbec2ebe3235e0904bac75cfbdc2c"} Dec 03 00:09:22 crc kubenswrapper[4953]: I1203 00:09:22.759164 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-t7ljb\" (UID: \"4313395c-8501-4b5b-a4cb-2fc986fced03\") " pod="openshift-image-registry/image-registry-697d97f7c8-t7ljb" Dec 03 00:09:22 crc kubenswrapper[4953]: E1203 00:09:22.760690 4953 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 00:09:23.260677872 +0000 UTC m=+152.244958655 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-t7ljb" (UID: "4313395c-8501-4b5b-a4cb-2fc986fced03") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 00:09:22 crc kubenswrapper[4953]: I1203 00:09:22.768103 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-8k8mf" event={"ID":"b2f23cca-bfc0-4e7a-bf15-2fc4065e840b","Type":"ContainerStarted","Data":"100fd5cc03f9cc20f5d8bf8f26a212c45f6cfe9491ac12834aabd4e2271e76b4"} Dec 03 00:09:22 crc kubenswrapper[4953]: I1203 00:09:22.769651 4953 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress/router-default-5444994796-r8l2l" podStartSLOduration=126.769638541 podStartE2EDuration="2m6.769638541s" podCreationTimestamp="2025-12-03 00:07:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 00:09:22.768381231 +0000 UTC m=+151.752662004" watchObservedRunningTime="2025-12-03 00:09:22.769638541 +0000 UTC m=+151.753919324" Dec 03 00:09:22 crc kubenswrapper[4953]: I1203 00:09:22.788440 4953 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-gdstj"] Dec 03 00:09:22 crc kubenswrapper[4953]: I1203 00:09:22.797041 4953 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-t968q"] Dec 03 00:09:22 crc kubenswrapper[4953]: I1203 00:09:22.804052 4953 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-xhc8f"] Dec 03 00:09:22 crc kubenswrapper[4953]: I1203 00:09:22.813752 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-gpwzf" event={"ID":"04f69b6e-9193-4055-8e50-e87fa663ca05","Type":"ContainerStarted","Data":"61c793e403beccdb173dbf64d1cc4dc5d26bf3f3db6fd98bca1e54156c8c45d5"} Dec 03 00:09:22 crc kubenswrapper[4953]: I1203 00:09:22.826187 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-l29cz" event={"ID":"2c280414-49b4-489f-b74a-13f6715744a6","Type":"ContainerStarted","Data":"5c5bc5074f8edb591d9e7661d380bbc234e0a9d2f34ca758dee5ff4567030ada"} Dec 03 00:09:22 crc kubenswrapper[4953]: I1203 00:09:22.843230 4953 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-server-gpwzf" podStartSLOduration=6.843212565 podStartE2EDuration="6.843212565s" podCreationTimestamp="2025-12-03 00:09:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 00:09:22.842307422 +0000 UTC m=+151.826588205" watchObservedRunningTime="2025-12-03 00:09:22.843212565 +0000 UTC m=+151.827493348" Dec 03 00:09:22 crc kubenswrapper[4953]: I1203 00:09:22.868407 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 00:09:22 crc kubenswrapper[4953]: E1203 00:09:22.868571 4953 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 00:09:23.368546455 +0000 UTC m=+152.352827228 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 00:09:22 crc kubenswrapper[4953]: I1203 00:09:22.868686 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-t7ljb\" (UID: \"4313395c-8501-4b5b-a4cb-2fc986fced03\") " pod="openshift-image-registry/image-registry-697d97f7c8-t7ljb" Dec 03 00:09:22 crc kubenswrapper[4953]: E1203 00:09:22.868949 4953 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 00:09:23.368939765 +0000 UTC m=+152.353220548 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-t7ljb" (UID: "4313395c-8501-4b5b-a4cb-2fc986fced03") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 00:09:22 crc kubenswrapper[4953]: W1203 00:09:22.918811 4953 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4da5668c_a644_4014_8358_790da0f313e2.slice/crio-2c8be04cb503e7a6678ab397d8e9c3d76f839f25a87e3fb9815758f1ce7f84c2 WatchSource:0}: Error finding container 2c8be04cb503e7a6678ab397d8e9c3d76f839f25a87e3fb9815758f1ce7f84c2: Status 404 returned error can't find the container with id 2c8be04cb503e7a6678ab397d8e9c3d76f839f25a87e3fb9815758f1ce7f84c2 Dec 03 00:09:22 crc kubenswrapper[4953]: I1203 00:09:22.952074 4953 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-z5glq"] Dec 03 00:09:23 crc kubenswrapper[4953]: I1203 00:09:22.969800 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 00:09:23 crc kubenswrapper[4953]: E1203 00:09:22.973460 4953 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 00:09:23.473432655 +0000 UTC m=+152.457713438 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 00:09:23 crc kubenswrapper[4953]: I1203 00:09:22.974404 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-t7ljb\" (UID: \"4313395c-8501-4b5b-a4cb-2fc986fced03\") " pod="openshift-image-registry/image-registry-697d97f7c8-t7ljb" Dec 03 00:09:23 crc kubenswrapper[4953]: E1203 00:09:22.977377 4953 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 00:09:23.477356252 +0000 UTC m=+152.461637035 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-t7ljb" (UID: "4313395c-8501-4b5b-a4cb-2fc986fced03") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 00:09:23 crc kubenswrapper[4953]: I1203 00:09:22.981372 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-xbwlm" event={"ID":"496793fd-c6ab-48e4-a2c5-c259593c5d1c","Type":"ContainerStarted","Data":"9662c55289fec9c4026634a16923736dc9acfaec59514dcf8def6fbee661407e"} Dec 03 00:09:23 crc kubenswrapper[4953]: I1203 00:09:22.990464 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-2bh27" event={"ID":"8c657b07-c109-4475-bdd9-94808f3a58a0","Type":"ContainerStarted","Data":"938ad8dee139af8f128f120f1f125590290360a9363db7d109730c57eb74babf"} Dec 03 00:09:23 crc kubenswrapper[4953]: I1203 00:09:22.995743 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-52hrr" event={"ID":"90247294-af78-4d04-8aed-af90f93a034a","Type":"ContainerStarted","Data":"5146c8aac8d7e3cb2d34f3059cb8805fa3cfd735cbdd42ef3ea7e72e404d7cce"} Dec 03 00:09:23 crc kubenswrapper[4953]: I1203 00:09:22.996225 4953 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-52hrr" Dec 03 00:09:23 crc kubenswrapper[4953]: I1203 00:09:23.043887 4953 patch_prober.go:28] interesting pod/route-controller-manager-6576b87f9c-52hrr container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.14:8443/healthz\": dial tcp 10.217.0.14:8443: connect: connection refused" start-of-body= Dec 03 00:09:23 crc kubenswrapper[4953]: I1203 00:09:23.043950 4953 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-52hrr" podUID="90247294-af78-4d04-8aed-af90f93a034a" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.14:8443/healthz\": dial tcp 10.217.0.14:8443: connect: connection refused" Dec 03 00:09:23 crc kubenswrapper[4953]: I1203 00:09:23.120857 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 00:09:23 crc kubenswrapper[4953]: E1203 00:09:23.130429 4953 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 00:09:23.630406442 +0000 UTC m=+152.614687215 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 00:09:23 crc kubenswrapper[4953]: I1203 00:09:23.139219 4953 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-llvsw"] Dec 03 00:09:23 crc kubenswrapper[4953]: I1203 00:09:23.140485 4953 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-zd9fs"] Dec 03 00:09:23 crc kubenswrapper[4953]: I1203 00:09:23.157094 4953 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-kx9dt"] Dec 03 00:09:23 crc kubenswrapper[4953]: I1203 00:09:23.173873 4953 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-l7m9z"] Dec 03 00:09:23 crc kubenswrapper[4953]: I1203 00:09:23.179486 4953 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-h95rc"] Dec 03 00:09:23 crc kubenswrapper[4953]: I1203 00:09:23.179524 4953 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-5fdxh"] Dec 03 00:09:23 crc kubenswrapper[4953]: I1203 00:09:23.181985 4953 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-m2msx"] Dec 03 00:09:23 crc kubenswrapper[4953]: I1203 00:09:23.209041 4953 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412000-mqgxv"] Dec 03 00:09:23 crc kubenswrapper[4953]: I1203 00:09:23.221193 4953 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-qv29t"] Dec 03 00:09:23 crc kubenswrapper[4953]: I1203 00:09:23.226024 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-t7ljb\" (UID: \"4313395c-8501-4b5b-a4cb-2fc986fced03\") " pod="openshift-image-registry/image-registry-697d97f7c8-t7ljb" Dec 03 00:09:23 crc kubenswrapper[4953]: E1203 00:09:23.226560 4953 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 00:09:23.726528477 +0000 UTC m=+152.710809260 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-t7ljb" (UID: "4313395c-8501-4b5b-a4cb-2fc986fced03") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 00:09:23 crc kubenswrapper[4953]: I1203 00:09:23.272223 4953 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-njtzc"] Dec 03 00:09:23 crc kubenswrapper[4953]: I1203 00:09:23.292667 4953 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-fcw6n"] Dec 03 00:09:23 crc kubenswrapper[4953]: I1203 00:09:23.312100 4953 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-2bh27" Dec 03 00:09:23 crc kubenswrapper[4953]: I1203 00:09:23.312370 4953 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-2bh27" Dec 03 00:09:23 crc kubenswrapper[4953]: I1203 00:09:23.340309 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 00:09:23 crc kubenswrapper[4953]: E1203 00:09:23.340709 4953 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 00:09:23.840694795 +0000 UTC m=+152.824975578 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 00:09:23 crc kubenswrapper[4953]: I1203 00:09:23.367678 4953 patch_prober.go:28] interesting pod/apiserver-7bbb656c7d-2bh27 container/oauth-apiserver namespace/openshift-oauth-apiserver: Startup probe status=failure output="Get \"https://10.217.0.5:8443/livez\": dial tcp 10.217.0.5:8443: connect: connection refused" start-of-body= Dec 03 00:09:23 crc kubenswrapper[4953]: I1203 00:09:23.367754 4953 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-2bh27" podUID="8c657b07-c109-4475-bdd9-94808f3a58a0" containerName="oauth-apiserver" probeResult="failure" output="Get \"https://10.217.0.5:8443/livez\": dial tcp 10.217.0.5:8443: connect: connection refused" Dec 03 00:09:23 crc kubenswrapper[4953]: I1203 00:09:23.368119 4953 patch_prober.go:28] interesting pod/router-default-5444994796-r8l2l container/router namespace/openshift-ingress: Startup probe status=failure output="Get \"http://localhost:1936/healthz/ready\": dial tcp [::1]:1936: connect: connection refused" start-of-body= Dec 03 00:09:23 crc kubenswrapper[4953]: I1203 00:09:23.368158 4953 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-r8l2l" podUID="4edc57f1-5c0c-4108-bb3c-b2f104c6e08b" containerName="router" probeResult="failure" output="Get \"http://localhost:1936/healthz/ready\": dial tcp [::1]:1936: connect: connection refused" Dec 03 00:09:23 crc kubenswrapper[4953]: I1203 00:09:23.387602 4953 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-2bh27" podStartSLOduration=127.387576294 podStartE2EDuration="2m7.387576294s" podCreationTimestamp="2025-12-03 00:07:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 00:09:23.339724862 +0000 UTC m=+152.324005645" watchObservedRunningTime="2025-12-03 00:09:23.387576294 +0000 UTC m=+152.371857087" Dec 03 00:09:23 crc kubenswrapper[4953]: I1203 00:09:23.403894 4953 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-ingress/router-default-5444994796-r8l2l" Dec 03 00:09:23 crc kubenswrapper[4953]: I1203 00:09:23.403928 4953 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-2xt2r"] Dec 03 00:09:23 crc kubenswrapper[4953]: I1203 00:09:23.424265 4953 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-brcxn"] Dec 03 00:09:23 crc kubenswrapper[4953]: I1203 00:09:23.435844 4953 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-kz9m9"] Dec 03 00:09:23 crc kubenswrapper[4953]: I1203 00:09:23.443440 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-t7ljb\" (UID: \"4313395c-8501-4b5b-a4cb-2fc986fced03\") " pod="openshift-image-registry/image-registry-697d97f7c8-t7ljb" Dec 03 00:09:23 crc kubenswrapper[4953]: E1203 00:09:23.443821 4953 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 00:09:23.943805292 +0000 UTC m=+152.928086075 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-t7ljb" (UID: "4313395c-8501-4b5b-a4cb-2fc986fced03") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 00:09:23 crc kubenswrapper[4953]: I1203 00:09:23.472650 4953 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-qbsjg"] Dec 03 00:09:23 crc kubenswrapper[4953]: I1203 00:09:23.534153 4953 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication-operator/authentication-operator-69f744f599-vcx7x" podStartSLOduration=128.534127835 podStartE2EDuration="2m8.534127835s" podCreationTimestamp="2025-12-03 00:07:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 00:09:23.485107384 +0000 UTC m=+152.469388177" watchObservedRunningTime="2025-12-03 00:09:23.534127835 +0000 UTC m=+152.518408618" Dec 03 00:09:23 crc kubenswrapper[4953]: I1203 00:09:23.534464 4953 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-jfp8j"] Dec 03 00:09:23 crc kubenswrapper[4953]: I1203 00:09:23.551753 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 00:09:23 crc kubenswrapper[4953]: E1203 00:09:23.552000 4953 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 00:09:24.051969313 +0000 UTC m=+153.036250086 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 00:09:23 crc kubenswrapper[4953]: I1203 00:09:23.578715 4953 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/machine-api-operator-5694c8668f-xbwlm" podStartSLOduration=127.578688017 podStartE2EDuration="2m7.578688017s" podCreationTimestamp="2025-12-03 00:07:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 00:09:23.537388705 +0000 UTC m=+152.521669488" watchObservedRunningTime="2025-12-03 00:09:23.578688017 +0000 UTC m=+152.562968800" Dec 03 00:09:23 crc kubenswrapper[4953]: I1203 00:09:23.659546 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-t7ljb\" (UID: \"4313395c-8501-4b5b-a4cb-2fc986fced03\") " pod="openshift-image-registry/image-registry-697d97f7c8-t7ljb" Dec 03 00:09:23 crc kubenswrapper[4953]: E1203 00:09:23.660182 4953 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 00:09:24.160164594 +0000 UTC m=+153.144445377 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-t7ljb" (UID: "4313395c-8501-4b5b-a4cb-2fc986fced03") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 00:09:24 crc kubenswrapper[4953]: I1203 00:09:23.760583 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 00:09:24 crc kubenswrapper[4953]: E1203 00:09:23.761201 4953 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 00:09:24.261185119 +0000 UTC m=+153.245465902 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 00:09:24 crc kubenswrapper[4953]: I1203 00:09:23.862102 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-t7ljb\" (UID: \"4313395c-8501-4b5b-a4cb-2fc986fced03\") " pod="openshift-image-registry/image-registry-697d97f7c8-t7ljb" Dec 03 00:09:24 crc kubenswrapper[4953]: E1203 00:09:23.862446 4953 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 00:09:24.362432361 +0000 UTC m=+153.346713144 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-t7ljb" (UID: "4313395c-8501-4b5b-a4cb-2fc986fced03") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 00:09:24 crc kubenswrapper[4953]: I1203 00:09:23.973049 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 00:09:24 crc kubenswrapper[4953]: E1203 00:09:23.973291 4953 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 00:09:24.473250436 +0000 UTC m=+153.457531219 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 00:09:24 crc kubenswrapper[4953]: I1203 00:09:23.973549 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-t7ljb\" (UID: \"4313395c-8501-4b5b-a4cb-2fc986fced03\") " pod="openshift-image-registry/image-registry-697d97f7c8-t7ljb" Dec 03 00:09:24 crc kubenswrapper[4953]: E1203 00:09:23.973908 4953 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 00:09:24.473897472 +0000 UTC m=+153.458178255 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-t7ljb" (UID: "4313395c-8501-4b5b-a4cb-2fc986fced03") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 00:09:24 crc kubenswrapper[4953]: I1203 00:09:24.050690 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-dlqdp" event={"ID":"536278ea-21b6-4193-ad9d-4dfb7ce6c194","Type":"ContainerStarted","Data":"35d0e8a97e05b294f13e7c36eed81a9fdb84e1440d015962e480a7dbac132568"} Dec 03 00:09:24 crc kubenswrapper[4953]: I1203 00:09:24.052753 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-m2msx" event={"ID":"4ce6dbe8-0292-4172-81a3-f5cdfdba6120","Type":"ContainerStarted","Data":"8e4400cc9926b1c11b9d3a84b70b88ad44403d4ea59a585783d40d47c3c61da7"} Dec 03 00:09:24 crc kubenswrapper[4953]: I1203 00:09:24.053637 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-t968q" event={"ID":"b77462cc-f0ef-4108-b81f-0d6d412d7fa8","Type":"ContainerStarted","Data":"ca9f2547249551901b1e9321282ddbd6f372013e4faa4a52343a56d451bcce11"} Dec 03 00:09:24 crc kubenswrapper[4953]: I1203 00:09:24.092661 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 00:09:24 crc kubenswrapper[4953]: E1203 00:09:24.093435 4953 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 00:09:24.593417481 +0000 UTC m=+153.577698264 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 00:09:24 crc kubenswrapper[4953]: I1203 00:09:24.122417 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-z5glq" event={"ID":"1ffff94d-cd60-4079-86d6-ff10eb285ca2","Type":"ContainerStarted","Data":"b5b9d125763fca1b82b1bf331da41ca585e77a83ff9647dee74ccb234a4ea57d"} Dec 03 00:09:24 crc kubenswrapper[4953]: I1203 00:09:24.197614 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-t7ljb\" (UID: \"4313395c-8501-4b5b-a4cb-2fc986fced03\") " pod="openshift-image-registry/image-registry-697d97f7c8-t7ljb" Dec 03 00:09:24 crc kubenswrapper[4953]: E1203 00:09:24.198328 4953 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 00:09:24.698302741 +0000 UTC m=+153.682583524 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-t7ljb" (UID: "4313395c-8501-4b5b-a4cb-2fc986fced03") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 00:09:24 crc kubenswrapper[4953]: I1203 00:09:24.235532 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-5fdxh" event={"ID":"912368ff-455d-4a2d-a444-08c54448ca2a","Type":"ContainerStarted","Data":"be0b2aa5c66732b8d6fb6c886bc22c922f0a9e1122ba4496f9973e5a918849ea"} Dec 03 00:09:24 crc kubenswrapper[4953]: I1203 00:09:24.247188 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-vcx7x" event={"ID":"6d99040d-450f-4ba8-a96d-16b694ca3636","Type":"ContainerStarted","Data":"cf1e83fd10ae0e9fa02acbddc0f25f0cacc219a7ab5243f2844160b31157fd31"} Dec 03 00:09:24 crc kubenswrapper[4953]: I1203 00:09:24.261885 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-zd9fs" event={"ID":"174354b2-d5cc-495a-ac13-cb56923aca08","Type":"ContainerStarted","Data":"17b5c33493ca4c8ae53d645d8f894a9a43e52760bdfb22faca8cae9d1f5c9c8e"} Dec 03 00:09:24 crc kubenswrapper[4953]: I1203 00:09:24.263627 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-lg8jg" event={"ID":"8a9efbc7-da01-4454-afce-64f004f240fd","Type":"ContainerStarted","Data":"314014a858e13a84ca2cdae62eeac1160839e6088d71fc74052363d4d25689e0"} Dec 03 00:09:24 crc kubenswrapper[4953]: I1203 00:09:24.263652 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-lg8jg" event={"ID":"8a9efbc7-da01-4454-afce-64f004f240fd","Type":"ContainerStarted","Data":"28a15a428c70a9d9def8cbfb704aa13597fa2e4f5083ed273d41a5a395f2c34e"} Dec 03 00:09:24 crc kubenswrapper[4953]: I1203 00:09:24.270181 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"6a84a068e20bcc7522ff0b1b5115000cf2599e0f2fc45bb37459bc165bb2a13a"} Dec 03 00:09:24 crc kubenswrapper[4953]: I1203 00:09:24.274375 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-lwq7b" event={"ID":"a7477307-0aa0-461d-8fc7-f4f8fc2f7c7b","Type":"ContainerStarted","Data":"c9250a7629298f80c41bba31f99619f3da04b9f62faa16b5b9e7e932f5c9ffd1"} Dec 03 00:09:24 crc kubenswrapper[4953]: I1203 00:09:24.277196 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-llvsw" event={"ID":"400790d0-9d4f-4698-bd68-8085ac29ec9c","Type":"ContainerStarted","Data":"363e8abac13e6adc381f1bf0a3ad8a492c0ac70500432d50cba50439b2c18d3c"} Dec 03 00:09:24 crc kubenswrapper[4953]: I1203 00:09:24.283561 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-gsszn" event={"ID":"416a5985-9d2e-43f0-be89-d8be0ae77816","Type":"ContainerStarted","Data":"479ea7a21d4c7bd717e4d6e7f47c39c77932588e37594e55a3287c3d867df1ed"} Dec 03 00:09:24 crc kubenswrapper[4953]: I1203 00:09:24.298330 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 00:09:24 crc kubenswrapper[4953]: E1203 00:09:24.298725 4953 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 00:09:24.798709901 +0000 UTC m=+153.782990684 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 00:09:24 crc kubenswrapper[4953]: I1203 00:09:24.305746 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-hzbcl" event={"ID":"466d51cd-18a2-419a-a749-1ab91c943f13","Type":"ContainerStarted","Data":"f477ec1bf5e2d8b22081b4775548149bdce6c367056749a52db9ce77bcbcc3e7"} Dec 03 00:09:24 crc kubenswrapper[4953]: I1203 00:09:24.319041 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-kx9dt" event={"ID":"7124b81d-b87c-45de-a112-20602c4d9adb","Type":"ContainerStarted","Data":"f8da00cbcdc757fed1f6c9775d6af01029548aca767571414b5aa799e2566daf"} Dec 03 00:09:24 crc kubenswrapper[4953]: I1203 00:09:24.342283 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-xhc8f" event={"ID":"4da5668c-a644-4014-8358-790da0f313e2","Type":"ContainerStarted","Data":"2c8be04cb503e7a6678ab397d8e9c3d76f839f25a87e3fb9815758f1ce7f84c2"} Dec 03 00:09:24 crc kubenswrapper[4953]: I1203 00:09:24.358645 4953 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-52hrr" podStartSLOduration=128.35862912 podStartE2EDuration="2m8.35862912s" podCreationTimestamp="2025-12-03 00:07:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 00:09:23.866319386 +0000 UTC m=+152.850600179" watchObservedRunningTime="2025-12-03 00:09:24.35862912 +0000 UTC m=+153.342909903" Dec 03 00:09:24 crc kubenswrapper[4953]: I1203 00:09:24.359834 4953 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-lg8jg" podStartSLOduration=128.359829839 podStartE2EDuration="2m8.359829839s" podCreationTimestamp="2025-12-03 00:07:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 00:09:24.357959304 +0000 UTC m=+153.342240087" watchObservedRunningTime="2025-12-03 00:09:24.359829839 +0000 UTC m=+153.344110622" Dec 03 00:09:24 crc kubenswrapper[4953]: I1203 00:09:24.364953 4953 patch_prober.go:28] interesting pod/router-default-5444994796-r8l2l container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 03 00:09:24 crc kubenswrapper[4953]: [-]has-synced failed: reason withheld Dec 03 00:09:24 crc kubenswrapper[4953]: [+]process-running ok Dec 03 00:09:24 crc kubenswrapper[4953]: healthz check failed Dec 03 00:09:24 crc kubenswrapper[4953]: I1203 00:09:24.365042 4953 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-r8l2l" podUID="4edc57f1-5c0c-4108-bb3c-b2f104c6e08b" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 03 00:09:24 crc kubenswrapper[4953]: I1203 00:09:24.401148 4953 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-lwq7b" podStartSLOduration=128.401135322 podStartE2EDuration="2m8.401135322s" podCreationTimestamp="2025-12-03 00:07:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 00:09:24.398719652 +0000 UTC m=+153.383000435" watchObservedRunningTime="2025-12-03 00:09:24.401135322 +0000 UTC m=+153.385416105" Dec 03 00:09:24 crc kubenswrapper[4953]: I1203 00:09:24.401314 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-t7ljb\" (UID: \"4313395c-8501-4b5b-a4cb-2fc986fced03\") " pod="openshift-image-registry/image-registry-697d97f7c8-t7ljb" Dec 03 00:09:24 crc kubenswrapper[4953]: E1203 00:09:24.403081 4953 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 00:09:24.903066349 +0000 UTC m=+153.887347202 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-t7ljb" (UID: "4313395c-8501-4b5b-a4cb-2fc986fced03") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 00:09:24 crc kubenswrapper[4953]: I1203 00:09:24.403736 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-gdstj" event={"ID":"6d44369d-4150-4832-9098-88d3c4b1133e","Type":"ContainerStarted","Data":"cdfb2109727301447b85a79797d045437985e0d8c6af72ef2a80db547fbd5fdf"} Dec 03 00:09:24 crc kubenswrapper[4953]: I1203 00:09:24.407910 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-lhvnr" event={"ID":"98788d99-bf86-41b8-b71b-0c1edab5dca2","Type":"ContainerStarted","Data":"c7ea3cccd4aebe7a15e4c2d4c491adfbc387274c8e3b70fdc257ef01da3032e2"} Dec 03 00:09:24 crc kubenswrapper[4953]: I1203 00:09:24.423080 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-49rsf" event={"ID":"0d6c7577-4632-4e4d-88f0-87548a6c640f","Type":"ContainerStarted","Data":"611bb5036436fe8ba0ffc76d1f9eeef9b47e542e7ce2432c87a765268cf13bcf"} Dec 03 00:09:24 crc kubenswrapper[4953]: I1203 00:09:24.502459 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 00:09:24 crc kubenswrapper[4953]: E1203 00:09:24.503879 4953 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 00:09:25.003862989 +0000 UTC m=+153.988143772 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 00:09:24 crc kubenswrapper[4953]: I1203 00:09:24.586588 4953 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-52hrr" Dec 03 00:09:24 crc kubenswrapper[4953]: I1203 00:09:24.602917 4953 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-lhvnr" podStartSLOduration=128.602902416 podStartE2EDuration="2m8.602902416s" podCreationTimestamp="2025-12-03 00:07:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 00:09:24.600723732 +0000 UTC m=+153.585004525" watchObservedRunningTime="2025-12-03 00:09:24.602902416 +0000 UTC m=+153.587183199" Dec 03 00:09:24 crc kubenswrapper[4953]: I1203 00:09:24.623127 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-t7ljb\" (UID: \"4313395c-8501-4b5b-a4cb-2fc986fced03\") " pod="openshift-image-registry/image-registry-697d97f7c8-t7ljb" Dec 03 00:09:24 crc kubenswrapper[4953]: E1203 00:09:24.623538 4953 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 00:09:25.123524181 +0000 UTC m=+154.107804964 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-t7ljb" (UID: "4313395c-8501-4b5b-a4cb-2fc986fced03") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 00:09:24 crc kubenswrapper[4953]: I1203 00:09:24.724460 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 00:09:24 crc kubenswrapper[4953]: E1203 00:09:24.724641 4953 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 00:09:25.224616178 +0000 UTC m=+154.208896961 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 00:09:24 crc kubenswrapper[4953]: I1203 00:09:24.724789 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-t7ljb\" (UID: \"4313395c-8501-4b5b-a4cb-2fc986fced03\") " pod="openshift-image-registry/image-registry-697d97f7c8-t7ljb" Dec 03 00:09:24 crc kubenswrapper[4953]: E1203 00:09:24.725147 4953 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 00:09:25.225134511 +0000 UTC m=+154.209415294 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-t7ljb" (UID: "4313395c-8501-4b5b-a4cb-2fc986fced03") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 00:09:24 crc kubenswrapper[4953]: I1203 00:09:24.826846 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 00:09:24 crc kubenswrapper[4953]: E1203 00:09:24.827059 4953 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 00:09:25.327028778 +0000 UTC m=+154.311309561 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 00:09:24 crc kubenswrapper[4953]: I1203 00:09:24.827998 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-t7ljb\" (UID: \"4313395c-8501-4b5b-a4cb-2fc986fced03\") " pod="openshift-image-registry/image-registry-697d97f7c8-t7ljb" Dec 03 00:09:24 crc kubenswrapper[4953]: E1203 00:09:24.828562 4953 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 00:09:25.328542035 +0000 UTC m=+154.312822818 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-t7ljb" (UID: "4313395c-8501-4b5b-a4cb-2fc986fced03") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 00:09:24 crc kubenswrapper[4953]: I1203 00:09:24.929450 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 00:09:24 crc kubenswrapper[4953]: E1203 00:09:24.929617 4953 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 00:09:25.429595011 +0000 UTC m=+154.413875804 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 00:09:24 crc kubenswrapper[4953]: I1203 00:09:24.929702 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-t7ljb\" (UID: \"4313395c-8501-4b5b-a4cb-2fc986fced03\") " pod="openshift-image-registry/image-registry-697d97f7c8-t7ljb" Dec 03 00:09:24 crc kubenswrapper[4953]: E1203 00:09:24.930080 4953 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 00:09:25.430069543 +0000 UTC m=+154.414350326 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-t7ljb" (UID: "4313395c-8501-4b5b-a4cb-2fc986fced03") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 00:09:25 crc kubenswrapper[4953]: I1203 00:09:25.049112 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 00:09:25 crc kubenswrapper[4953]: E1203 00:09:25.049349 4953 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 00:09:25.549326045 +0000 UTC m=+154.533606828 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 00:09:25 crc kubenswrapper[4953]: I1203 00:09:25.150282 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-t7ljb\" (UID: \"4313395c-8501-4b5b-a4cb-2fc986fced03\") " pod="openshift-image-registry/image-registry-697d97f7c8-t7ljb" Dec 03 00:09:25 crc kubenswrapper[4953]: E1203 00:09:25.150673 4953 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 00:09:25.650661099 +0000 UTC m=+154.634941882 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-t7ljb" (UID: "4313395c-8501-4b5b-a4cb-2fc986fced03") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 00:09:25 crc kubenswrapper[4953]: I1203 00:09:25.292182 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 00:09:25 crc kubenswrapper[4953]: E1203 00:09:25.292314 4953 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 00:09:25.792276969 +0000 UTC m=+154.776557772 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 00:09:25 crc kubenswrapper[4953]: I1203 00:09:25.296006 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-t7ljb\" (UID: \"4313395c-8501-4b5b-a4cb-2fc986fced03\") " pod="openshift-image-registry/image-registry-697d97f7c8-t7ljb" Dec 03 00:09:25 crc kubenswrapper[4953]: E1203 00:09:25.296537 4953 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 00:09:25.796485822 +0000 UTC m=+154.780766605 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-t7ljb" (UID: "4313395c-8501-4b5b-a4cb-2fc986fced03") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 00:09:25 crc kubenswrapper[4953]: I1203 00:09:25.366295 4953 patch_prober.go:28] interesting pod/router-default-5444994796-r8l2l container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 03 00:09:25 crc kubenswrapper[4953]: [-]has-synced failed: reason withheld Dec 03 00:09:25 crc kubenswrapper[4953]: [+]process-running ok Dec 03 00:09:25 crc kubenswrapper[4953]: healthz check failed Dec 03 00:09:25 crc kubenswrapper[4953]: I1203 00:09:25.366574 4953 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-r8l2l" podUID="4edc57f1-5c0c-4108-bb3c-b2f104c6e08b" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 03 00:09:25 crc kubenswrapper[4953]: I1203 00:09:25.397610 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 00:09:25 crc kubenswrapper[4953]: E1203 00:09:25.398015 4953 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 00:09:25.89799839 +0000 UTC m=+154.882279173 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 00:09:25 crc kubenswrapper[4953]: I1203 00:09:25.512392 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-t7ljb\" (UID: \"4313395c-8501-4b5b-a4cb-2fc986fced03\") " pod="openshift-image-registry/image-registry-697d97f7c8-t7ljb" Dec 03 00:09:25 crc kubenswrapper[4953]: E1203 00:09:25.513005 4953 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 00:09:26.012993618 +0000 UTC m=+154.997274401 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-t7ljb" (UID: "4313395c-8501-4b5b-a4cb-2fc986fced03") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 00:09:25 crc kubenswrapper[4953]: I1203 00:09:25.613471 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 00:09:25 crc kubenswrapper[4953]: E1203 00:09:25.613771 4953 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 00:09:26.113754677 +0000 UTC m=+155.098035460 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 00:09:25 crc kubenswrapper[4953]: I1203 00:09:25.673287 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-zd9fs" event={"ID":"174354b2-d5cc-495a-ac13-cb56923aca08","Type":"ContainerStarted","Data":"bd00ff0885618566b92f2391187e9d3672332ed049ecbb80a2e5bd769ca9d084"} Dec 03 00:09:25 crc kubenswrapper[4953]: I1203 00:09:25.732863 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-t7ljb\" (UID: \"4313395c-8501-4b5b-a4cb-2fc986fced03\") " pod="openshift-image-registry/image-registry-697d97f7c8-t7ljb" Dec 03 00:09:25 crc kubenswrapper[4953]: E1203 00:09:25.733208 4953 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 00:09:26.233194204 +0000 UTC m=+155.217474987 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-t7ljb" (UID: "4313395c-8501-4b5b-a4cb-2fc986fced03") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 00:09:25 crc kubenswrapper[4953]: I1203 00:09:25.769151 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-95vxq" event={"ID":"a8a759e4-551c-41f9-bbab-9cfbf490777a","Type":"ContainerStarted","Data":"e97e886657617662ab7ef27bb5e80215dd15a426e9ed8e80c7758da05f8cca7c"} Dec 03 00:09:25 crc kubenswrapper[4953]: I1203 00:09:25.794411 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-llvsw" event={"ID":"400790d0-9d4f-4698-bd68-8085ac29ec9c","Type":"ContainerStarted","Data":"a4d729c2b26493f505c618b9dc1507ac6753fbd6b1827c263025825c8f136f5f"} Dec 03 00:09:25 crc kubenswrapper[4953]: I1203 00:09:25.825686 4953 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca/service-ca-9c57cc56f-zd9fs" podStartSLOduration=129.82567028 podStartE2EDuration="2m9.82567028s" podCreationTimestamp="2025-12-03 00:07:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 00:09:25.82075824 +0000 UTC m=+154.805039023" watchObservedRunningTime="2025-12-03 00:09:25.82567028 +0000 UTC m=+154.809951063" Dec 03 00:09:25 crc kubenswrapper[4953]: I1203 00:09:25.834355 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 00:09:25 crc kubenswrapper[4953]: I1203 00:09:25.831854 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-z5glq" event={"ID":"1ffff94d-cd60-4079-86d6-ff10eb285ca2","Type":"ContainerStarted","Data":"c682f97979aa4f6ba919c0addb99684f145d79f190c67c26dccd1eb6429d678f"} Dec 03 00:09:25 crc kubenswrapper[4953]: E1203 00:09:25.835213 4953 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 00:09:26.335186563 +0000 UTC m=+155.319467346 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 00:09:25 crc kubenswrapper[4953]: I1203 00:09:25.838600 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-t7ljb\" (UID: \"4313395c-8501-4b5b-a4cb-2fc986fced03\") " pod="openshift-image-registry/image-registry-697d97f7c8-t7ljb" Dec 03 00:09:25 crc kubenswrapper[4953]: E1203 00:09:25.839156 4953 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 00:09:26.3391451 +0000 UTC m=+155.323425883 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-t7ljb" (UID: "4313395c-8501-4b5b-a4cb-2fc986fced03") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 00:09:25 crc kubenswrapper[4953]: I1203 00:09:25.850588 4953 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-canary/ingress-canary-llvsw" podStartSLOduration=9.850572999 podStartE2EDuration="9.850572999s" podCreationTimestamp="2025-12-03 00:09:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 00:09:25.8497467 +0000 UTC m=+154.834027483" watchObservedRunningTime="2025-12-03 00:09:25.850572999 +0000 UTC m=+154.834853782" Dec 03 00:09:25 crc kubenswrapper[4953]: I1203 00:09:25.857107 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-8k8mf" event={"ID":"b2f23cca-bfc0-4e7a-bf15-2fc4065e840b","Type":"ContainerStarted","Data":"4f9e32839310d39888c6073451106900913fd506f4c9cf5b63b31f75453ad167"} Dec 03 00:09:25 crc kubenswrapper[4953]: I1203 00:09:25.868996 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-qv29t" event={"ID":"c893caae-db46-4210-8be1-0ccf40bc9913","Type":"ContainerStarted","Data":"94d364b3cc9298381e8eb4ed47d1a609588b6bf569cc1fa2e9168615ee709c5e"} Dec 03 00:09:25 crc kubenswrapper[4953]: I1203 00:09:25.898543 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-gsszn" event={"ID":"416a5985-9d2e-43f0-be89-d8be0ae77816","Type":"ContainerStarted","Data":"6150bb0a68356a2349778d62bcba34ac9263ca7ff7c0073a082d7c6e901bf72c"} Dec 03 00:09:25 crc kubenswrapper[4953]: I1203 00:09:25.914397 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-t968q" event={"ID":"b77462cc-f0ef-4108-b81f-0d6d412d7fa8","Type":"ContainerStarted","Data":"f8d2ab30665fb0bf1e755513ddd7dec64764708ea5b10e426cef653d78cca96b"} Dec 03 00:09:25 crc kubenswrapper[4953]: I1203 00:09:25.916656 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-fcw6n" event={"ID":"16c23c64-3f6c-4ce9-9118-cfeb890dfeed","Type":"ContainerStarted","Data":"8c4f13272f510ae700628057b2ac6ecbd86dfc793445c394c29c07a51ab8602b"} Dec 03 00:09:25 crc kubenswrapper[4953]: I1203 00:09:25.916682 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-fcw6n" event={"ID":"16c23c64-3f6c-4ce9-9118-cfeb890dfeed","Type":"ContainerStarted","Data":"fa96d4affa8142bfda12d7b5f980e9df21135dfad93442d262883f285b2a68cb"} Dec 03 00:09:25 crc kubenswrapper[4953]: I1203 00:09:25.918855 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-njtzc" event={"ID":"fc4d1764-e8db-415b-9650-ed9a03cc49d1","Type":"ContainerStarted","Data":"a1dbc012f13e3757f2c086fd53bd67788098fcec7fb75e04f10fb3e05062c96b"} Dec 03 00:09:25 crc kubenswrapper[4953]: I1203 00:09:25.918882 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-njtzc" event={"ID":"fc4d1764-e8db-415b-9650-ed9a03cc49d1","Type":"ContainerStarted","Data":"ce58917f95af844136fa4f2e6b9724478cb67133ac102316e7d668dd647b246e"} Dec 03 00:09:25 crc kubenswrapper[4953]: I1203 00:09:25.940793 4953 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-95vxq" podStartSLOduration=129.940766549 podStartE2EDuration="2m9.940766549s" podCreationTimestamp="2025-12-03 00:07:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 00:09:25.932074956 +0000 UTC m=+154.916355739" watchObservedRunningTime="2025-12-03 00:09:25.940766549 +0000 UTC m=+154.925047332" Dec 03 00:09:25 crc kubenswrapper[4953]: I1203 00:09:25.942744 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 00:09:25 crc kubenswrapper[4953]: E1203 00:09:25.944159 4953 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 00:09:26.444142992 +0000 UTC m=+155.428423775 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 00:09:25 crc kubenswrapper[4953]: I1203 00:09:25.973696 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-brcxn" event={"ID":"1ee4772d-f4e3-4da8-a520-805dc2317fdd","Type":"ContainerStarted","Data":"5c523f546d2081551798e24312d77129fc310366936d2e5e2be39f713309402c"} Dec 03 00:09:25 crc kubenswrapper[4953]: I1203 00:09:25.973744 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-brcxn" event={"ID":"1ee4772d-f4e3-4da8-a520-805dc2317fdd","Type":"ContainerStarted","Data":"3cc7399f0743a6c28467ea52a3404f11920a03729a0abda561943c6b5cf3a2fc"} Dec 03 00:09:25 crc kubenswrapper[4953]: I1203 00:09:25.975459 4953 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-brcxn" Dec 03 00:09:26 crc kubenswrapper[4953]: I1203 00:09:26.012248 4953 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-brcxn container/catalog-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.24:8443/healthz\": dial tcp 10.217.0.24:8443: connect: connection refused" start-of-body= Dec 03 00:09:26 crc kubenswrapper[4953]: I1203 00:09:26.012345 4953 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-brcxn" podUID="1ee4772d-f4e3-4da8-a520-805dc2317fdd" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.24:8443/healthz\": dial tcp 10.217.0.24:8443: connect: connection refused" Dec 03 00:09:26 crc kubenswrapper[4953]: I1203 00:09:26.013324 4953 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-njtzc" podStartSLOduration=130.013301157 podStartE2EDuration="2m10.013301157s" podCreationTimestamp="2025-12-03 00:07:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 00:09:25.967807022 +0000 UTC m=+154.952087805" watchObservedRunningTime="2025-12-03 00:09:26.013301157 +0000 UTC m=+154.997581940" Dec 03 00:09:26 crc kubenswrapper[4953]: I1203 00:09:26.013597 4953 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns-operator/dns-operator-744455d44c-8k8mf" podStartSLOduration=130.013592574 podStartE2EDuration="2m10.013592574s" podCreationTimestamp="2025-12-03 00:07:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 00:09:26.01341398 +0000 UTC m=+154.997694763" watchObservedRunningTime="2025-12-03 00:09:26.013592574 +0000 UTC m=+154.997873357" Dec 03 00:09:26 crc kubenswrapper[4953]: I1203 00:09:26.022478 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-jfp8j" event={"ID":"13c8b8e5-a48a-4376-be38-f4902b55ad94","Type":"ContainerStarted","Data":"3c148acd4e09e139508487364ab033a2797688bc51e85526810a26d6ac4ffde1"} Dec 03 00:09:26 crc kubenswrapper[4953]: I1203 00:09:26.046305 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-t7ljb\" (UID: \"4313395c-8501-4b5b-a4cb-2fc986fced03\") " pod="openshift-image-registry/image-registry-697d97f7c8-t7ljb" Dec 03 00:09:26 crc kubenswrapper[4953]: E1203 00:09:26.048778 4953 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 00:09:26.548762486 +0000 UTC m=+155.533043269 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-t7ljb" (UID: "4313395c-8501-4b5b-a4cb-2fc986fced03") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 00:09:26 crc kubenswrapper[4953]: I1203 00:09:26.077028 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-xhc8f" event={"ID":"4da5668c-a644-4014-8358-790da0f313e2","Type":"ContainerStarted","Data":"25659d3612fac67500d9cef49d5331ae72cc96369cfd5774132584115cd4b307"} Dec 03 00:09:26 crc kubenswrapper[4953]: I1203 00:09:26.149090 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 00:09:26 crc kubenswrapper[4953]: E1203 00:09:26.150309 4953 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 00:09:26.650289764 +0000 UTC m=+155.634570547 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 00:09:26 crc kubenswrapper[4953]: I1203 00:09:26.173074 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-hzbcl" event={"ID":"466d51cd-18a2-419a-a749-1ab91c943f13","Type":"ContainerStarted","Data":"e420b4615d55d796bc1f16c5f7a7b20e461fdbc388b5e1f477c31eafad02859a"} Dec 03 00:09:26 crc kubenswrapper[4953]: I1203 00:09:26.175172 4953 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-admission-controller-857f4d67dd-gsszn" podStartSLOduration=130.175159773 podStartE2EDuration="2m10.175159773s" podCreationTimestamp="2025-12-03 00:07:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 00:09:26.08200667 +0000 UTC m=+155.066287473" watchObservedRunningTime="2025-12-03 00:09:26.175159773 +0000 UTC m=+155.159440556" Dec 03 00:09:26 crc kubenswrapper[4953]: I1203 00:09:26.175569 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-l29cz" event={"ID":"2c280414-49b4-489f-b74a-13f6715744a6","Type":"ContainerStarted","Data":"0895119f38370023cd8605267541ccf9c70b56d51c3c1eafb18f8d0bfd349816"} Dec 03 00:09:26 crc kubenswrapper[4953]: I1203 00:09:26.176874 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-2xt2r" event={"ID":"73033ee8-cae5-4acc-8c77-60905eed62f2","Type":"ContainerStarted","Data":"491837959c820ad9ad527e2684a39054b17b56860df1c5c7d98ec6769d351dd5"} Dec 03 00:09:26 crc kubenswrapper[4953]: I1203 00:09:26.178620 4953 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd-operator/etcd-operator-b45778765-t968q" podStartSLOduration=130.178609738 podStartE2EDuration="2m10.178609738s" podCreationTimestamp="2025-12-03 00:07:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 00:09:26.174601189 +0000 UTC m=+155.158881972" watchObservedRunningTime="2025-12-03 00:09:26.178609738 +0000 UTC m=+155.162890521" Dec 03 00:09:26 crc kubenswrapper[4953]: I1203 00:09:26.205730 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-lhvnr" event={"ID":"98788d99-bf86-41b8-b71b-0c1edab5dca2","Type":"ContainerStarted","Data":"dc6c27b527524d0e26b4f4b4dc3ecf75a71af9b3ce7810128a78c69d9fc3c946"} Dec 03 00:09:26 crc kubenswrapper[4953]: I1203 00:09:26.245703 4953 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-gdstj" podStartSLOduration=131.245672391 podStartE2EDuration="2m11.245672391s" podCreationTimestamp="2025-12-03 00:07:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 00:09:26.245494897 +0000 UTC m=+155.229775680" watchObservedRunningTime="2025-12-03 00:09:26.245672391 +0000 UTC m=+155.229953174" Dec 03 00:09:26 crc kubenswrapper[4953]: I1203 00:09:26.248280 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-5fdxh" event={"ID":"912368ff-455d-4a2d-a444-08c54448ca2a","Type":"ContainerStarted","Data":"8c5b8318ba51bcc60fa074a4166b8689a6f02b9c4db5eedda3850b44147fc6bb"} Dec 03 00:09:26 crc kubenswrapper[4953]: I1203 00:09:26.251450 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-t7ljb\" (UID: \"4313395c-8501-4b5b-a4cb-2fc986fced03\") " pod="openshift-image-registry/image-registry-697d97f7c8-t7ljb" Dec 03 00:09:26 crc kubenswrapper[4953]: E1203 00:09:26.252878 4953 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 00:09:26.752858897 +0000 UTC m=+155.737139870 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-t7ljb" (UID: "4313395c-8501-4b5b-a4cb-2fc986fced03") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 00:09:26 crc kubenswrapper[4953]: I1203 00:09:26.264538 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-qbsjg" event={"ID":"1f9a783f-9a9e-47b6-bc01-9c25261b1f2b","Type":"ContainerStarted","Data":"d9be763849658803d97403f16330411e6a6609100fea28c06bdab8ff9d3c4dee"} Dec 03 00:09:26 crc kubenswrapper[4953]: I1203 00:09:26.313739 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29412000-mqgxv" event={"ID":"7b91526b-2492-4c17-949b-0ba3f8c221ea","Type":"ContainerStarted","Data":"c9bb8e9e994fc6c10d6aab1c55ae333228332a4ed3369558948c3249ba7c71af"} Dec 03 00:09:26 crc kubenswrapper[4953]: I1203 00:09:26.313797 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29412000-mqgxv" event={"ID":"7b91526b-2492-4c17-949b-0ba3f8c221ea","Type":"ContainerStarted","Data":"cb0ad0b7aefd5f2be56b688771208a14b485d7dfd5b9b676113f11e5b27a7302"} Dec 03 00:09:26 crc kubenswrapper[4953]: I1203 00:09:26.316194 4953 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-z5glq" podStartSLOduration=130.316183559 podStartE2EDuration="2m10.316183559s" podCreationTimestamp="2025-12-03 00:07:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 00:09:26.288825229 +0000 UTC m=+155.273106012" watchObservedRunningTime="2025-12-03 00:09:26.316183559 +0000 UTC m=+155.300464342" Dec 03 00:09:26 crc kubenswrapper[4953]: I1203 00:09:26.333249 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-kz9m9" event={"ID":"48b0efc4-5e8a-4bd9-9016-26bdd4761bf1","Type":"ContainerStarted","Data":"c7c7f0392bd05852fad2fe62f98fcf0ef5ba5a87efc7e1ce3c0f598aa22629c9"} Dec 03 00:09:26 crc kubenswrapper[4953]: I1203 00:09:26.333310 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-kz9m9" event={"ID":"48b0efc4-5e8a-4bd9-9016-26bdd4761bf1","Type":"ContainerStarted","Data":"2258114029a0fb5865bf1f083d14fd9e5f2f87f38c58b4244c356f6c9931f005"} Dec 03 00:09:26 crc kubenswrapper[4953]: I1203 00:09:26.359209 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 00:09:26 crc kubenswrapper[4953]: E1203 00:09:26.360269 4953 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 00:09:26.860248599 +0000 UTC m=+155.844529382 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 00:09:26 crc kubenswrapper[4953]: I1203 00:09:26.371107 4953 patch_prober.go:28] interesting pod/router-default-5444994796-r8l2l container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 03 00:09:26 crc kubenswrapper[4953]: [-]has-synced failed: reason withheld Dec 03 00:09:26 crc kubenswrapper[4953]: [+]process-running ok Dec 03 00:09:26 crc kubenswrapper[4953]: healthz check failed Dec 03 00:09:26 crc kubenswrapper[4953]: I1203 00:09:26.371161 4953 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-r8l2l" podUID="4edc57f1-5c0c-4108-bb3c-b2f104c6e08b" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 03 00:09:26 crc kubenswrapper[4953]: I1203 00:09:26.393930 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-49rsf" event={"ID":"0d6c7577-4632-4e4d-88f0-87548a6c640f","Type":"ContainerStarted","Data":"c0a5b771854e53b21ff398c326cabf8a517f4703a1cdf709fae6c3d17c6a3a5e"} Dec 03 00:09:26 crc kubenswrapper[4953]: I1203 00:09:26.402081 4953 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-l29cz" podStartSLOduration=130.402066714 podStartE2EDuration="2m10.402066714s" podCreationTimestamp="2025-12-03 00:07:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 00:09:26.401413838 +0000 UTC m=+155.385694611" watchObservedRunningTime="2025-12-03 00:09:26.402066714 +0000 UTC m=+155.386347497" Dec 03 00:09:26 crc kubenswrapper[4953]: I1203 00:09:26.403451 4953 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca-operator/service-ca-operator-777779d784-fcw6n" podStartSLOduration=130.403444397 podStartE2EDuration="2m10.403444397s" podCreationTimestamp="2025-12-03 00:07:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 00:09:26.317334007 +0000 UTC m=+155.301614790" watchObservedRunningTime="2025-12-03 00:09:26.403444397 +0000 UTC m=+155.387725180" Dec 03 00:09:26 crc kubenswrapper[4953]: I1203 00:09:26.440245 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-l7m9z" event={"ID":"2000e3b0-f8e7-4f0a-9f59-8c9d73324586","Type":"ContainerStarted","Data":"e0eac16b66e178fc34fb566f6a184be24fafe1275aa315358db4e2df1a7d4c8e"} Dec 03 00:09:26 crc kubenswrapper[4953]: I1203 00:09:26.459502 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-dlqdp" event={"ID":"536278ea-21b6-4193-ad9d-4dfb7ce6c194","Type":"ContainerStarted","Data":"f8427e3b2b460f050dfdd8480f097ceb7a2e52f8978a5e34d4aa365578528b2b"} Dec 03 00:09:26 crc kubenswrapper[4953]: I1203 00:09:26.460078 4953 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-dlqdp" Dec 03 00:09:26 crc kubenswrapper[4953]: I1203 00:09:26.460224 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-t7ljb\" (UID: \"4313395c-8501-4b5b-a4cb-2fc986fced03\") " pod="openshift-image-registry/image-registry-697d97f7c8-t7ljb" Dec 03 00:09:26 crc kubenswrapper[4953]: E1203 00:09:26.460959 4953 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 00:09:26.960948277 +0000 UTC m=+155.945229060 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-t7ljb" (UID: "4313395c-8501-4b5b-a4cb-2fc986fced03") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 00:09:26 crc kubenswrapper[4953]: I1203 00:09:26.466991 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-h95rc" event={"ID":"e7e8210f-8915-49c3-8e95-f4427962f66a","Type":"ContainerStarted","Data":"b31aa89261eddf14aaf3b92693282e9b7ce0cc580ebd49daef3c12ae2aa73723"} Dec 03 00:09:26 crc kubenswrapper[4953]: I1203 00:09:26.467022 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-h95rc" event={"ID":"e7e8210f-8915-49c3-8e95-f4427962f66a","Type":"ContainerStarted","Data":"19beec69429f1c33d6feed6a4a2b451bd82c7a1aad36cb26c1dc5530e530a284"} Dec 03 00:09:26 crc kubenswrapper[4953]: I1203 00:09:26.467036 4953 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-h95rc" Dec 03 00:09:26 crc kubenswrapper[4953]: I1203 00:09:26.480701 4953 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-h95rc container/olm-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.36:8443/healthz\": dial tcp 10.217.0.36:8443: connect: connection refused" start-of-body= Dec 03 00:09:26 crc kubenswrapper[4953]: I1203 00:09:26.480747 4953 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-h95rc" podUID="e7e8210f-8915-49c3-8e95-f4427962f66a" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.36:8443/healthz\": dial tcp 10.217.0.36:8443: connect: connection refused" Dec 03 00:09:26 crc kubenswrapper[4953]: I1203 00:09:26.496560 4953 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-xhc8f" podStartSLOduration=130.496542239 podStartE2EDuration="2m10.496542239s" podCreationTimestamp="2025-12-03 00:07:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 00:09:26.495161505 +0000 UTC m=+155.479442288" watchObservedRunningTime="2025-12-03 00:09:26.496542239 +0000 UTC m=+155.480823022" Dec 03 00:09:26 crc kubenswrapper[4953]: I1203 00:09:26.527882 4953 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-5fdxh" podStartSLOduration=130.527863296 podStartE2EDuration="2m10.527863296s" podCreationTimestamp="2025-12-03 00:07:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 00:09:26.527519578 +0000 UTC m=+155.511800361" watchObservedRunningTime="2025-12-03 00:09:26.527863296 +0000 UTC m=+155.512144079" Dec 03 00:09:26 crc kubenswrapper[4953]: I1203 00:09:26.561506 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 00:09:26 crc kubenswrapper[4953]: E1203 00:09:26.561713 4953 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 00:09:27.061680445 +0000 UTC m=+156.045961238 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 00:09:26 crc kubenswrapper[4953]: I1203 00:09:26.562176 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-t7ljb\" (UID: \"4313395c-8501-4b5b-a4cb-2fc986fced03\") " pod="openshift-image-registry/image-registry-697d97f7c8-t7ljb" Dec 03 00:09:26 crc kubenswrapper[4953]: E1203 00:09:26.564648 4953 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 00:09:27.064635748 +0000 UTC m=+156.048916531 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-t7ljb" (UID: "4313395c-8501-4b5b-a4cb-2fc986fced03") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 00:09:26 crc kubenswrapper[4953]: I1203 00:09:26.576297 4953 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29412000-mqgxv" podStartSLOduration=131.576282003 podStartE2EDuration="2m11.576282003s" podCreationTimestamp="2025-12-03 00:07:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 00:09:26.57452562 +0000 UTC m=+155.558806403" watchObservedRunningTime="2025-12-03 00:09:26.576282003 +0000 UTC m=+155.560562786" Dec 03 00:09:26 crc kubenswrapper[4953]: I1203 00:09:26.613434 4953 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-brcxn" podStartSLOduration=130.613415733 podStartE2EDuration="2m10.613415733s" podCreationTimestamp="2025-12-03 00:07:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 00:09:26.612948621 +0000 UTC m=+155.597229404" watchObservedRunningTime="2025-12-03 00:09:26.613415733 +0000 UTC m=+155.597696516" Dec 03 00:09:26 crc kubenswrapper[4953]: I1203 00:09:26.663199 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 00:09:26 crc kubenswrapper[4953]: E1203 00:09:26.664041 4953 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 00:09:27.164021623 +0000 UTC m=+156.148302406 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 00:09:26 crc kubenswrapper[4953]: I1203 00:09:26.667070 4953 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver/apiserver-76f77b778f-hzbcl" podStartSLOduration=131.667044257 podStartE2EDuration="2m11.667044257s" podCreationTimestamp="2025-12-03 00:07:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 00:09:26.666374861 +0000 UTC m=+155.650655644" watchObservedRunningTime="2025-12-03 00:09:26.667044257 +0000 UTC m=+155.651325040" Dec 03 00:09:26 crc kubenswrapper[4953]: I1203 00:09:26.748820 4953 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-dlqdp" podStartSLOduration=130.748805771 podStartE2EDuration="2m10.748805771s" podCreationTimestamp="2025-12-03 00:07:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 00:09:26.739026001 +0000 UTC m=+155.723306784" watchObservedRunningTime="2025-12-03 00:09:26.748805771 +0000 UTC m=+155.733086554" Dec 03 00:09:26 crc kubenswrapper[4953]: I1203 00:09:26.764708 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-t7ljb\" (UID: \"4313395c-8501-4b5b-a4cb-2fc986fced03\") " pod="openshift-image-registry/image-registry-697d97f7c8-t7ljb" Dec 03 00:09:26 crc kubenswrapper[4953]: E1203 00:09:26.764967 4953 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 00:09:27.264955396 +0000 UTC m=+156.249236179 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-t7ljb" (UID: "4313395c-8501-4b5b-a4cb-2fc986fced03") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 00:09:26 crc kubenswrapper[4953]: I1203 00:09:26.866368 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 00:09:26 crc kubenswrapper[4953]: E1203 00:09:26.866713 4953 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 00:09:27.36669809 +0000 UTC m=+156.350978873 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 00:09:26 crc kubenswrapper[4953]: I1203 00:09:26.967288 4953 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-kz9m9" podStartSLOduration=130.967267374 podStartE2EDuration="2m10.967267374s" podCreationTimestamp="2025-12-03 00:07:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 00:09:26.965156762 +0000 UTC m=+155.949437545" watchObservedRunningTime="2025-12-03 00:09:26.967267374 +0000 UTC m=+155.951548157" Dec 03 00:09:26 crc kubenswrapper[4953]: I1203 00:09:26.968012 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-t7ljb\" (UID: \"4313395c-8501-4b5b-a4cb-2fc986fced03\") " pod="openshift-image-registry/image-registry-697d97f7c8-t7ljb" Dec 03 00:09:26 crc kubenswrapper[4953]: E1203 00:09:26.968293 4953 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 00:09:27.468282529 +0000 UTC m=+156.452563302 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-t7ljb" (UID: "4313395c-8501-4b5b-a4cb-2fc986fced03") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 00:09:27 crc kubenswrapper[4953]: I1203 00:09:27.060348 4953 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-h95rc" podStartSLOduration=131.060332514 podStartE2EDuration="2m11.060332514s" podCreationTimestamp="2025-12-03 00:07:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 00:09:27.060313094 +0000 UTC m=+156.044593877" watchObservedRunningTime="2025-12-03 00:09:27.060332514 +0000 UTC m=+156.044613297" Dec 03 00:09:27 crc kubenswrapper[4953]: I1203 00:09:27.073379 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 00:09:27 crc kubenswrapper[4953]: E1203 00:09:27.073655 4953 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 00:09:27.57359412 +0000 UTC m=+156.557874903 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 00:09:27 crc kubenswrapper[4953]: I1203 00:09:27.073814 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-t7ljb\" (UID: \"4313395c-8501-4b5b-a4cb-2fc986fced03\") " pod="openshift-image-registry/image-registry-697d97f7c8-t7ljb" Dec 03 00:09:27 crc kubenswrapper[4953]: E1203 00:09:27.074377 4953 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 00:09:27.574368849 +0000 UTC m=+156.558649632 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-t7ljb" (UID: "4313395c-8501-4b5b-a4cb-2fc986fced03") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 00:09:27 crc kubenswrapper[4953]: I1203 00:09:27.091568 4953 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-49rsf" podStartSLOduration=131.091550619 podStartE2EDuration="2m11.091550619s" podCreationTimestamp="2025-12-03 00:07:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 00:09:27.090104734 +0000 UTC m=+156.074385517" watchObservedRunningTime="2025-12-03 00:09:27.091550619 +0000 UTC m=+156.075831402" Dec 03 00:09:27 crc kubenswrapper[4953]: I1203 00:09:27.175616 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 00:09:27 crc kubenswrapper[4953]: E1203 00:09:27.175959 4953 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 00:09:27.675935957 +0000 UTC m=+156.660216740 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 00:09:27 crc kubenswrapper[4953]: I1203 00:09:27.277707 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-t7ljb\" (UID: \"4313395c-8501-4b5b-a4cb-2fc986fced03\") " pod="openshift-image-registry/image-registry-697d97f7c8-t7ljb" Dec 03 00:09:27 crc kubenswrapper[4953]: E1203 00:09:27.278050 4953 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 00:09:27.778038469 +0000 UTC m=+156.762319252 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-t7ljb" (UID: "4313395c-8501-4b5b-a4cb-2fc986fced03") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 00:09:27 crc kubenswrapper[4953]: I1203 00:09:27.357888 4953 patch_prober.go:28] interesting pod/router-default-5444994796-r8l2l container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 03 00:09:27 crc kubenswrapper[4953]: [-]has-synced failed: reason withheld Dec 03 00:09:27 crc kubenswrapper[4953]: [+]process-running ok Dec 03 00:09:27 crc kubenswrapper[4953]: healthz check failed Dec 03 00:09:27 crc kubenswrapper[4953]: I1203 00:09:27.358277 4953 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-r8l2l" podUID="4edc57f1-5c0c-4108-bb3c-b2f104c6e08b" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 03 00:09:27 crc kubenswrapper[4953]: I1203 00:09:27.379103 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 00:09:27 crc kubenswrapper[4953]: E1203 00:09:27.379474 4953 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 00:09:27.879456745 +0000 UTC m=+156.863737528 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 00:09:27 crc kubenswrapper[4953]: I1203 00:09:27.480082 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-t7ljb\" (UID: \"4313395c-8501-4b5b-a4cb-2fc986fced03\") " pod="openshift-image-registry/image-registry-697d97f7c8-t7ljb" Dec 03 00:09:27 crc kubenswrapper[4953]: E1203 00:09:27.480401 4953 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 00:09:27.980385328 +0000 UTC m=+156.964666111 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-t7ljb" (UID: "4313395c-8501-4b5b-a4cb-2fc986fced03") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 00:09:27 crc kubenswrapper[4953]: I1203 00:09:27.504696 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-2xt2r" event={"ID":"73033ee8-cae5-4acc-8c77-60905eed62f2","Type":"ContainerStarted","Data":"be732f8169aaa92f92c5c61e404f9209c01d6df98afda00c50d525c27503343b"} Dec 03 00:09:27 crc kubenswrapper[4953]: I1203 00:09:27.580790 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 00:09:27 crc kubenswrapper[4953]: E1203 00:09:27.581067 4953 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 00:09:28.081052815 +0000 UTC m=+157.065333598 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 00:09:27 crc kubenswrapper[4953]: I1203 00:09:27.588767 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-5fdxh" event={"ID":"912368ff-455d-4a2d-a444-08c54448ca2a","Type":"ContainerStarted","Data":"dfe4b8fdb9829f85006c2375dfde8b54782e5c190856cadfc71134ee4d28dacb"} Dec 03 00:09:27 crc kubenswrapper[4953]: I1203 00:09:27.590613 4953 generic.go:334] "Generic (PLEG): container finished" podID="2000e3b0-f8e7-4f0a-9f59-8c9d73324586" containerID="a968e092313175ab6fc4cc2506c82b032c2f3c51318394a99ff9255cb1f55d54" exitCode=0 Dec 03 00:09:27 crc kubenswrapper[4953]: I1203 00:09:27.590648 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-l7m9z" event={"ID":"2000e3b0-f8e7-4f0a-9f59-8c9d73324586","Type":"ContainerStarted","Data":"5b8ab7e2f028d6c783ddf0ee53b6f7426d00cc4df2785025f042b3a390462fc2"} Dec 03 00:09:27 crc kubenswrapper[4953]: I1203 00:09:27.590661 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-l7m9z" event={"ID":"2000e3b0-f8e7-4f0a-9f59-8c9d73324586","Type":"ContainerDied","Data":"a968e092313175ab6fc4cc2506c82b032c2f3c51318394a99ff9255cb1f55d54"} Dec 03 00:09:27 crc kubenswrapper[4953]: I1203 00:09:27.591340 4953 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-config-operator/openshift-config-operator-7777fb866f-l7m9z" Dec 03 00:09:27 crc kubenswrapper[4953]: I1203 00:09:27.592339 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-qv29t" event={"ID":"c893caae-db46-4210-8be1-0ccf40bc9913","Type":"ContainerStarted","Data":"f491ababd60b69ce887fb2c21722305f9e556a9ff5e2fa1da22a017ac1c7a0f0"} Dec 03 00:09:27 crc kubenswrapper[4953]: I1203 00:09:27.592998 4953 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-qv29t" Dec 03 00:09:27 crc kubenswrapper[4953]: I1203 00:09:27.593837 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-kx9dt" event={"ID":"7124b81d-b87c-45de-a112-20602c4d9adb","Type":"ContainerStarted","Data":"cf37583f0764dfdbf86f5e9f0dae7b270a80910c7793261d30e36b161b46cea5"} Dec 03 00:09:27 crc kubenswrapper[4953]: I1203 00:09:27.594357 4953 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-kx9dt" Dec 03 00:09:27 crc kubenswrapper[4953]: I1203 00:09:27.595221 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-m2msx" event={"ID":"4ce6dbe8-0292-4172-81a3-f5cdfdba6120","Type":"ContainerStarted","Data":"8d0fbdf0b418b88a82cda2a76160b8ce4d6b2f600d6e2836c3c0469cc39e77e8"} Dec 03 00:09:27 crc kubenswrapper[4953]: I1203 00:09:27.596705 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-qbsjg" event={"ID":"1f9a783f-9a9e-47b6-bc01-9c25261b1f2b","Type":"ContainerStarted","Data":"9edaa68524999148dc028e90d48b06d050ff4989f59eea9028b215c29baae470"} Dec 03 00:09:27 crc kubenswrapper[4953]: I1203 00:09:27.596727 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-qbsjg" event={"ID":"1f9a783f-9a9e-47b6-bc01-9c25261b1f2b","Type":"ContainerStarted","Data":"8ec9cd82c4ac0973a2092b1a0acec6e6a17f540b2c64f402bb44a0ea7e3d6d89"} Dec 03 00:09:27 crc kubenswrapper[4953]: I1203 00:09:27.597136 4953 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-dns/dns-default-qbsjg" Dec 03 00:09:27 crc kubenswrapper[4953]: I1203 00:09:27.598159 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-jfp8j" event={"ID":"13c8b8e5-a48a-4376-be38-f4902b55ad94","Type":"ContainerStarted","Data":"1b8ac01725b64da1cce1d5123d7bdbe55476eac3c7e10f87c132f7b0854a3e8b"} Dec 03 00:09:27 crc kubenswrapper[4953]: I1203 00:09:27.599585 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-gdstj" event={"ID":"6d44369d-4150-4832-9098-88d3c4b1133e","Type":"ContainerStarted","Data":"569d6a0a594e2700e571ed220c46ed5728f57bfee1fba58d13b5ce8e1084db2e"} Dec 03 00:09:27 crc kubenswrapper[4953]: I1203 00:09:27.615843 4953 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-kx9dt container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.27:8080/healthz\": dial tcp 10.217.0.27:8080: connect: connection refused" start-of-body= Dec 03 00:09:27 crc kubenswrapper[4953]: I1203 00:09:27.615995 4953 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-kx9dt" podUID="7124b81d-b87c-45de-a112-20602c4d9adb" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.27:8080/healthz\": dial tcp 10.217.0.27:8080: connect: connection refused" Dec 03 00:09:27 crc kubenswrapper[4953]: I1203 00:09:27.634208 4953 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-h95rc" Dec 03 00:09:27 crc kubenswrapper[4953]: I1203 00:09:27.652157 4953 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-config-operator/openshift-config-operator-7777fb866f-l7m9z" podStartSLOduration=131.652136137 podStartE2EDuration="2m11.652136137s" podCreationTimestamp="2025-12-03 00:07:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 00:09:27.651798788 +0000 UTC m=+156.636079591" watchObservedRunningTime="2025-12-03 00:09:27.652136137 +0000 UTC m=+156.636416930" Dec 03 00:09:27 crc kubenswrapper[4953]: I1203 00:09:27.670332 4953 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-qv29t" podStartSLOduration=131.670313292 podStartE2EDuration="2m11.670313292s" podCreationTimestamp="2025-12-03 00:07:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 00:09:27.670154048 +0000 UTC m=+156.654434831" watchObservedRunningTime="2025-12-03 00:09:27.670313292 +0000 UTC m=+156.654594075" Dec 03 00:09:27 crc kubenswrapper[4953]: I1203 00:09:27.679249 4953 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-brcxn" Dec 03 00:09:27 crc kubenswrapper[4953]: I1203 00:09:27.684579 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-t7ljb\" (UID: \"4313395c-8501-4b5b-a4cb-2fc986fced03\") " pod="openshift-image-registry/image-registry-697d97f7c8-t7ljb" Dec 03 00:09:27 crc kubenswrapper[4953]: E1203 00:09:27.697089 4953 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 00:09:28.197072218 +0000 UTC m=+157.181352991 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-t7ljb" (UID: "4313395c-8501-4b5b-a4cb-2fc986fced03") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 00:09:27 crc kubenswrapper[4953]: I1203 00:09:27.708215 4953 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-kx9dt" podStartSLOduration=131.708199901 podStartE2EDuration="2m11.708199901s" podCreationTimestamp="2025-12-03 00:07:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 00:09:27.707413001 +0000 UTC m=+156.691693784" watchObservedRunningTime="2025-12-03 00:09:27.708199901 +0000 UTC m=+156.692480684" Dec 03 00:09:27 crc kubenswrapper[4953]: I1203 00:09:27.738848 4953 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/dns-default-qbsjg" podStartSLOduration=11.738826461 podStartE2EDuration="11.738826461s" podCreationTimestamp="2025-12-03 00:09:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 00:09:27.737261343 +0000 UTC m=+156.721542126" watchObservedRunningTime="2025-12-03 00:09:27.738826461 +0000 UTC m=+156.723107244" Dec 03 00:09:27 crc kubenswrapper[4953]: I1203 00:09:27.786485 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 00:09:27 crc kubenswrapper[4953]: E1203 00:09:27.786829 4953 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 00:09:28.286803267 +0000 UTC m=+157.271084050 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 00:09:27 crc kubenswrapper[4953]: I1203 00:09:27.786871 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-t7ljb\" (UID: \"4313395c-8501-4b5b-a4cb-2fc986fced03\") " pod="openshift-image-registry/image-registry-697d97f7c8-t7ljb" Dec 03 00:09:27 crc kubenswrapper[4953]: E1203 00:09:27.787320 4953 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 00:09:28.287312889 +0000 UTC m=+157.271593672 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-t7ljb" (UID: "4313395c-8501-4b5b-a4cb-2fc986fced03") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 00:09:27 crc kubenswrapper[4953]: I1203 00:09:27.857567 4953 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-jfp8j" podStartSLOduration=131.857552941 podStartE2EDuration="2m11.857552941s" podCreationTimestamp="2025-12-03 00:07:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 00:09:27.818918664 +0000 UTC m=+156.803199447" watchObservedRunningTime="2025-12-03 00:09:27.857552941 +0000 UTC m=+156.841833724" Dec 03 00:09:27 crc kubenswrapper[4953]: I1203 00:09:27.858417 4953 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-f9d7485db-m2msx" podStartSLOduration=131.858412962 podStartE2EDuration="2m11.858412962s" podCreationTimestamp="2025-12-03 00:07:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 00:09:27.856993487 +0000 UTC m=+156.841274290" watchObservedRunningTime="2025-12-03 00:09:27.858412962 +0000 UTC m=+156.842693745" Dec 03 00:09:27 crc kubenswrapper[4953]: I1203 00:09:27.887391 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 00:09:27 crc kubenswrapper[4953]: E1203 00:09:27.887782 4953 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 00:09:28.387749591 +0000 UTC m=+157.372030374 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 00:09:27 crc kubenswrapper[4953]: I1203 00:09:27.988688 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-t7ljb\" (UID: \"4313395c-8501-4b5b-a4cb-2fc986fced03\") " pod="openshift-image-registry/image-registry-697d97f7c8-t7ljb" Dec 03 00:09:27 crc kubenswrapper[4953]: E1203 00:09:27.989115 4953 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 00:09:28.488982141 +0000 UTC m=+157.473262924 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-t7ljb" (UID: "4313395c-8501-4b5b-a4cb-2fc986fced03") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 00:09:28 crc kubenswrapper[4953]: I1203 00:09:28.098823 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 00:09:28 crc kubenswrapper[4953]: E1203 00:09:28.099014 4953 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 00:09:28.598986437 +0000 UTC m=+157.583267220 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 00:09:28 crc kubenswrapper[4953]: I1203 00:09:28.099216 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-t7ljb\" (UID: \"4313395c-8501-4b5b-a4cb-2fc986fced03\") " pod="openshift-image-registry/image-registry-697d97f7c8-t7ljb" Dec 03 00:09:28 crc kubenswrapper[4953]: E1203 00:09:28.099594 4953 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 00:09:28.599576721 +0000 UTC m=+157.583857494 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-t7ljb" (UID: "4313395c-8501-4b5b-a4cb-2fc986fced03") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 00:09:28 crc kubenswrapper[4953]: I1203 00:09:28.198043 4953 patch_prober.go:28] interesting pod/downloads-7954f5f757-5rzxr container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" start-of-body= Dec 03 00:09:28 crc kubenswrapper[4953]: I1203 00:09:28.198408 4953 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-5rzxr" podUID="e0e3abc1-91d2-4a6a-aa1c-dc166f30d75a" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" Dec 03 00:09:28 crc kubenswrapper[4953]: I1203 00:09:28.198157 4953 patch_prober.go:28] interesting pod/downloads-7954f5f757-5rzxr container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" start-of-body= Dec 03 00:09:28 crc kubenswrapper[4953]: I1203 00:09:28.198479 4953 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-5rzxr" podUID="e0e3abc1-91d2-4a6a-aa1c-dc166f30d75a" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" Dec 03 00:09:28 crc kubenswrapper[4953]: I1203 00:09:28.201003 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 00:09:28 crc kubenswrapper[4953]: E1203 00:09:28.201378 4953 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 00:09:28.701356825 +0000 UTC m=+157.685637608 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 00:09:28 crc kubenswrapper[4953]: I1203 00:09:28.326142 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-t7ljb\" (UID: \"4313395c-8501-4b5b-a4cb-2fc986fced03\") " pod="openshift-image-registry/image-registry-697d97f7c8-t7ljb" Dec 03 00:09:28 crc kubenswrapper[4953]: E1203 00:09:28.326769 4953 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 00:09:28.826749828 +0000 UTC m=+157.811030611 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-t7ljb" (UID: "4313395c-8501-4b5b-a4cb-2fc986fced03") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 00:09:28 crc kubenswrapper[4953]: I1203 00:09:28.333783 4953 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-2bh27" Dec 03 00:09:28 crc kubenswrapper[4953]: I1203 00:09:28.339305 4953 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-2bh27" Dec 03 00:09:28 crc kubenswrapper[4953]: I1203 00:09:28.353278 4953 patch_prober.go:28] interesting pod/router-default-5444994796-r8l2l container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 03 00:09:28 crc kubenswrapper[4953]: [-]has-synced failed: reason withheld Dec 03 00:09:28 crc kubenswrapper[4953]: [+]process-running ok Dec 03 00:09:28 crc kubenswrapper[4953]: healthz check failed Dec 03 00:09:28 crc kubenswrapper[4953]: I1203 00:09:28.353354 4953 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-r8l2l" podUID="4edc57f1-5c0c-4108-bb3c-b2f104c6e08b" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 03 00:09:28 crc kubenswrapper[4953]: I1203 00:09:28.423064 4953 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console-operator/console-operator-58897d9998-c56d6" Dec 03 00:09:28 crc kubenswrapper[4953]: I1203 00:09:28.427723 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 00:09:28 crc kubenswrapper[4953]: E1203 00:09:28.429112 4953 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 00:09:28.929081036 +0000 UTC m=+157.913361819 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 00:09:28 crc kubenswrapper[4953]: I1203 00:09:28.495617 4953 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-apiserver/apiserver-76f77b778f-hzbcl" Dec 03 00:09:28 crc kubenswrapper[4953]: I1203 00:09:28.495675 4953 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-apiserver/apiserver-76f77b778f-hzbcl" Dec 03 00:09:28 crc kubenswrapper[4953]: I1203 00:09:28.529454 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-t7ljb\" (UID: \"4313395c-8501-4b5b-a4cb-2fc986fced03\") " pod="openshift-image-registry/image-registry-697d97f7c8-t7ljb" Dec 03 00:09:28 crc kubenswrapper[4953]: E1203 00:09:28.531414 4953 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 00:09:29.031392703 +0000 UTC m=+158.015673486 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-t7ljb" (UID: "4313395c-8501-4b5b-a4cb-2fc986fced03") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 00:09:28 crc kubenswrapper[4953]: I1203 00:09:28.593417 4953 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-qv29t container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.32:5443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 03 00:09:28 crc kubenswrapper[4953]: I1203 00:09:28.593492 4953 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-qv29t" podUID="c893caae-db46-4210-8be1-0ccf40bc9913" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.32:5443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 03 00:09:28 crc kubenswrapper[4953]: I1203 00:09:28.614434 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-2xt2r" event={"ID":"73033ee8-cae5-4acc-8c77-60905eed62f2","Type":"ContainerStarted","Data":"b795dcb8798783e279eb3ca310a28f4c854e5c0c6e20d65bd708b5be2d83582d"} Dec 03 00:09:28 crc kubenswrapper[4953]: I1203 00:09:28.618218 4953 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-kx9dt container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.27:8080/healthz\": dial tcp 10.217.0.27:8080: connect: connection refused" start-of-body= Dec 03 00:09:28 crc kubenswrapper[4953]: I1203 00:09:28.618287 4953 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-kx9dt" podUID="7124b81d-b87c-45de-a112-20602c4d9adb" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.27:8080/healthz\": dial tcp 10.217.0.27:8080: connect: connection refused" Dec 03 00:09:28 crc kubenswrapper[4953]: I1203 00:09:28.631577 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 00:09:28 crc kubenswrapper[4953]: E1203 00:09:28.631897 4953 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 00:09:29.131880665 +0000 UTC m=+158.116161448 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 00:09:28 crc kubenswrapper[4953]: I1203 00:09:28.733995 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-t7ljb\" (UID: \"4313395c-8501-4b5b-a4cb-2fc986fced03\") " pod="openshift-image-registry/image-registry-697d97f7c8-t7ljb" Dec 03 00:09:28 crc kubenswrapper[4953]: E1203 00:09:28.735184 4953 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 00:09:29.235166146 +0000 UTC m=+158.219446929 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-t7ljb" (UID: "4313395c-8501-4b5b-a4cb-2fc986fced03") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 00:09:28 crc kubenswrapper[4953]: I1203 00:09:28.802519 4953 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-qv29t" Dec 03 00:09:28 crc kubenswrapper[4953]: I1203 00:09:28.836194 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 00:09:28 crc kubenswrapper[4953]: E1203 00:09:28.836443 4953 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 00:09:29.336412467 +0000 UTC m=+158.320693250 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 00:09:28 crc kubenswrapper[4953]: I1203 00:09:28.836811 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-t7ljb\" (UID: \"4313395c-8501-4b5b-a4cb-2fc986fced03\") " pod="openshift-image-registry/image-registry-697d97f7c8-t7ljb" Dec 03 00:09:28 crc kubenswrapper[4953]: E1203 00:09:28.837409 4953 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 00:09:29.337391131 +0000 UTC m=+158.321671914 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-t7ljb" (UID: "4313395c-8501-4b5b-a4cb-2fc986fced03") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 00:09:28 crc kubenswrapper[4953]: I1203 00:09:28.940016 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 00:09:28 crc kubenswrapper[4953]: E1203 00:09:28.941199 4953 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 00:09:29.441179865 +0000 UTC m=+158.425460648 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 00:09:29 crc kubenswrapper[4953]: I1203 00:09:29.050267 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-t7ljb\" (UID: \"4313395c-8501-4b5b-a4cb-2fc986fced03\") " pod="openshift-image-registry/image-registry-697d97f7c8-t7ljb" Dec 03 00:09:29 crc kubenswrapper[4953]: E1203 00:09:29.050667 4953 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 00:09:29.550648297 +0000 UTC m=+158.534929080 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-t7ljb" (UID: "4313395c-8501-4b5b-a4cb-2fc986fced03") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 00:09:29 crc kubenswrapper[4953]: I1203 00:09:29.152506 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 00:09:29 crc kubenswrapper[4953]: E1203 00:09:29.152772 4953 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 00:09:29.652756669 +0000 UTC m=+158.637037452 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 00:09:29 crc kubenswrapper[4953]: I1203 00:09:29.200955 4953 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-f9d7485db-m2msx" Dec 03 00:09:29 crc kubenswrapper[4953]: I1203 00:09:29.201087 4953 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-f9d7485db-m2msx" Dec 03 00:09:29 crc kubenswrapper[4953]: I1203 00:09:29.209461 4953 patch_prober.go:28] interesting pod/console-f9d7485db-m2msx container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.16:8443/health\": dial tcp 10.217.0.16:8443: connect: connection refused" start-of-body= Dec 03 00:09:29 crc kubenswrapper[4953]: I1203 00:09:29.209529 4953 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-m2msx" podUID="4ce6dbe8-0292-4172-81a3-f5cdfdba6120" containerName="console" probeResult="failure" output="Get \"https://10.217.0.16:8443/health\": dial tcp 10.217.0.16:8443: connect: connection refused" Dec 03 00:09:29 crc kubenswrapper[4953]: I1203 00:09:29.253670 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-t7ljb\" (UID: \"4313395c-8501-4b5b-a4cb-2fc986fced03\") " pod="openshift-image-registry/image-registry-697d97f7c8-t7ljb" Dec 03 00:09:29 crc kubenswrapper[4953]: E1203 00:09:29.254105 4953 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 00:09:29.754083562 +0000 UTC m=+158.738364415 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-t7ljb" (UID: "4313395c-8501-4b5b-a4cb-2fc986fced03") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 00:09:29 crc kubenswrapper[4953]: I1203 00:09:29.352254 4953 patch_prober.go:28] interesting pod/router-default-5444994796-r8l2l container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 03 00:09:29 crc kubenswrapper[4953]: [-]has-synced failed: reason withheld Dec 03 00:09:29 crc kubenswrapper[4953]: [+]process-running ok Dec 03 00:09:29 crc kubenswrapper[4953]: healthz check failed Dec 03 00:09:29 crc kubenswrapper[4953]: I1203 00:09:29.352363 4953 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-r8l2l" podUID="4edc57f1-5c0c-4108-bb3c-b2f104c6e08b" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 03 00:09:29 crc kubenswrapper[4953]: I1203 00:09:29.354345 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 00:09:29 crc kubenswrapper[4953]: E1203 00:09:29.354684 4953 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 00:09:29.854666567 +0000 UTC m=+158.838947350 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 00:09:29 crc kubenswrapper[4953]: I1203 00:09:29.355293 4953 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ingress/router-default-5444994796-r8l2l" Dec 03 00:09:29 crc kubenswrapper[4953]: I1203 00:09:29.459449 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-t7ljb\" (UID: \"4313395c-8501-4b5b-a4cb-2fc986fced03\") " pod="openshift-image-registry/image-registry-697d97f7c8-t7ljb" Dec 03 00:09:29 crc kubenswrapper[4953]: E1203 00:09:29.460138 4953 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 00:09:29.96010515 +0000 UTC m=+158.944386113 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-t7ljb" (UID: "4313395c-8501-4b5b-a4cb-2fc986fced03") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 00:09:29 crc kubenswrapper[4953]: I1203 00:09:29.560359 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 00:09:29 crc kubenswrapper[4953]: E1203 00:09:29.560730 4953 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 00:09:30.060714725 +0000 UTC m=+159.044995508 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 00:09:29 crc kubenswrapper[4953]: I1203 00:09:29.650751 4953 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-kx9dt container/marketplace-operator namespace/openshift-marketplace: Liveness probe status=failure output="Get \"http://10.217.0.27:8080/healthz\": dial tcp 10.217.0.27:8080: connect: connection refused" start-of-body= Dec 03 00:09:29 crc kubenswrapper[4953]: I1203 00:09:29.650853 4953 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-marketplace/marketplace-operator-79b997595-kx9dt" podUID="7124b81d-b87c-45de-a112-20602c4d9adb" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.27:8080/healthz\": dial tcp 10.217.0.27:8080: connect: connection refused" Dec 03 00:09:29 crc kubenswrapper[4953]: I1203 00:09:29.650754 4953 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-kx9dt container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.27:8080/healthz\": dial tcp 10.217.0.27:8080: connect: connection refused" start-of-body= Dec 03 00:09:29 crc kubenswrapper[4953]: I1203 00:09:29.651143 4953 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-kx9dt" podUID="7124b81d-b87c-45de-a112-20602c4d9adb" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.27:8080/healthz\": dial tcp 10.217.0.27:8080: connect: connection refused" Dec 03 00:09:29 crc kubenswrapper[4953]: I1203 00:09:29.663864 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-t7ljb\" (UID: \"4313395c-8501-4b5b-a4cb-2fc986fced03\") " pod="openshift-image-registry/image-registry-697d97f7c8-t7ljb" Dec 03 00:09:29 crc kubenswrapper[4953]: E1203 00:09:29.664299 4953 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 00:09:30.164283253 +0000 UTC m=+159.148564036 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-t7ljb" (UID: "4313395c-8501-4b5b-a4cb-2fc986fced03") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 00:09:29 crc kubenswrapper[4953]: I1203 00:09:29.665204 4953 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-kx9dt container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.27:8080/healthz\": dial tcp 10.217.0.27:8080: connect: connection refused" start-of-body= Dec 03 00:09:29 crc kubenswrapper[4953]: I1203 00:09:29.665281 4953 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-kx9dt" podUID="7124b81d-b87c-45de-a112-20602c4d9adb" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.27:8080/healthz\": dial tcp 10.217.0.27:8080: connect: connection refused" Dec 03 00:09:29 crc kubenswrapper[4953]: I1203 00:09:29.765200 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 00:09:29 crc kubenswrapper[4953]: E1203 00:09:29.766774 4953 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 00:09:30.266750014 +0000 UTC m=+159.251030797 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 00:09:29 crc kubenswrapper[4953]: I1203 00:09:29.855684 4953 patch_prober.go:28] interesting pod/apiserver-76f77b778f-hzbcl container/openshift-apiserver namespace/openshift-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[+]ping ok Dec 03 00:09:29 crc kubenswrapper[4953]: [+]log ok Dec 03 00:09:29 crc kubenswrapper[4953]: [+]etcd ok Dec 03 00:09:29 crc kubenswrapper[4953]: [+]poststarthook/start-apiserver-admission-initializer ok Dec 03 00:09:29 crc kubenswrapper[4953]: [+]poststarthook/generic-apiserver-start-informers ok Dec 03 00:09:29 crc kubenswrapper[4953]: [+]poststarthook/max-in-flight-filter ok Dec 03 00:09:29 crc kubenswrapper[4953]: [+]poststarthook/storage-object-count-tracker-hook ok Dec 03 00:09:29 crc kubenswrapper[4953]: [+]poststarthook/image.openshift.io-apiserver-caches ok Dec 03 00:09:29 crc kubenswrapper[4953]: [-]poststarthook/authorization.openshift.io-bootstrapclusterroles failed: reason withheld Dec 03 00:09:29 crc kubenswrapper[4953]: [-]poststarthook/authorization.openshift.io-ensurenodebootstrap-sa failed: reason withheld Dec 03 00:09:29 crc kubenswrapper[4953]: [+]poststarthook/project.openshift.io-projectcache ok Dec 03 00:09:29 crc kubenswrapper[4953]: [+]poststarthook/project.openshift.io-projectauthorizationcache ok Dec 03 00:09:29 crc kubenswrapper[4953]: [-]poststarthook/openshift.io-startinformers failed: reason withheld Dec 03 00:09:29 crc kubenswrapper[4953]: [+]poststarthook/openshift.io-restmapperupdater ok Dec 03 00:09:29 crc kubenswrapper[4953]: [+]poststarthook/quota.openshift.io-clusterquotamapping ok Dec 03 00:09:29 crc kubenswrapper[4953]: livez check failed Dec 03 00:09:29 crc kubenswrapper[4953]: I1203 00:09:29.855785 4953 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-apiserver/apiserver-76f77b778f-hzbcl" podUID="466d51cd-18a2-419a-a749-1ab91c943f13" containerName="openshift-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 03 00:09:29 crc kubenswrapper[4953]: I1203 00:09:29.867358 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-t7ljb\" (UID: \"4313395c-8501-4b5b-a4cb-2fc986fced03\") " pod="openshift-image-registry/image-registry-697d97f7c8-t7ljb" Dec 03 00:09:29 crc kubenswrapper[4953]: E1203 00:09:29.867932 4953 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 00:09:30.367909463 +0000 UTC m=+159.352190246 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-t7ljb" (UID: "4313395c-8501-4b5b-a4cb-2fc986fced03") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 00:09:29 crc kubenswrapper[4953]: I1203 00:09:29.878348 4953 plugin_watcher.go:194] "Adding socket path or updating timestamp to desired state cache" path="/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock" Dec 03 00:09:29 crc kubenswrapper[4953]: I1203 00:09:29.974608 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 00:09:29 crc kubenswrapper[4953]: E1203 00:09:29.975626 4953 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 00:09:30.475595302 +0000 UTC m=+159.459876085 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 00:09:30 crc kubenswrapper[4953]: I1203 00:09:30.049938 4953 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-hs6pg"] Dec 03 00:09:30 crc kubenswrapper[4953]: I1203 00:09:30.070948 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-hs6pg" Dec 03 00:09:30 crc kubenswrapper[4953]: I1203 00:09:30.070826 4953 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-xfd2l"] Dec 03 00:09:30 crc kubenswrapper[4953]: I1203 00:09:30.073556 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-xfd2l" Dec 03 00:09:30 crc kubenswrapper[4953]: I1203 00:09:30.074468 4953 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Dec 03 00:09:30 crc kubenswrapper[4953]: I1203 00:09:30.077703 4953 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-xfd2l"] Dec 03 00:09:30 crc kubenswrapper[4953]: I1203 00:09:30.080643 4953 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Dec 03 00:09:30 crc kubenswrapper[4953]: I1203 00:09:30.081926 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-t7ljb\" (UID: \"4313395c-8501-4b5b-a4cb-2fc986fced03\") " pod="openshift-image-registry/image-registry-697d97f7c8-t7ljb" Dec 03 00:09:30 crc kubenswrapper[4953]: E1203 00:09:30.082401 4953 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 00:09:30.582378909 +0000 UTC m=+159.566659692 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-t7ljb" (UID: "4313395c-8501-4b5b-a4cb-2fc986fced03") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 00:09:30 crc kubenswrapper[4953]: I1203 00:09:30.102346 4953 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-hs6pg"] Dec 03 00:09:30 crc kubenswrapper[4953]: I1203 00:09:30.183115 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 00:09:30 crc kubenswrapper[4953]: I1203 00:09:30.183339 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qlmtg\" (UniqueName: \"kubernetes.io/projected/52fd5773-3888-49f1-813d-10d262f4c901-kube-api-access-qlmtg\") pod \"certified-operators-xfd2l\" (UID: \"52fd5773-3888-49f1-813d-10d262f4c901\") " pod="openshift-marketplace/certified-operators-xfd2l" Dec 03 00:09:30 crc kubenswrapper[4953]: I1203 00:09:30.183388 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/52fd5773-3888-49f1-813d-10d262f4c901-utilities\") pod \"certified-operators-xfd2l\" (UID: \"52fd5773-3888-49f1-813d-10d262f4c901\") " pod="openshift-marketplace/certified-operators-xfd2l" Dec 03 00:09:30 crc kubenswrapper[4953]: I1203 00:09:30.183410 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m89m9\" (UniqueName: \"kubernetes.io/projected/6b353eef-b7f9-4dbc-b2aa-9860220438cc-kube-api-access-m89m9\") pod \"community-operators-hs6pg\" (UID: \"6b353eef-b7f9-4dbc-b2aa-9860220438cc\") " pod="openshift-marketplace/community-operators-hs6pg" Dec 03 00:09:30 crc kubenswrapper[4953]: I1203 00:09:30.183456 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/52fd5773-3888-49f1-813d-10d262f4c901-catalog-content\") pod \"certified-operators-xfd2l\" (UID: \"52fd5773-3888-49f1-813d-10d262f4c901\") " pod="openshift-marketplace/certified-operators-xfd2l" Dec 03 00:09:30 crc kubenswrapper[4953]: I1203 00:09:30.183491 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6b353eef-b7f9-4dbc-b2aa-9860220438cc-utilities\") pod \"community-operators-hs6pg\" (UID: \"6b353eef-b7f9-4dbc-b2aa-9860220438cc\") " pod="openshift-marketplace/community-operators-hs6pg" Dec 03 00:09:30 crc kubenswrapper[4953]: I1203 00:09:30.183594 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6b353eef-b7f9-4dbc-b2aa-9860220438cc-catalog-content\") pod \"community-operators-hs6pg\" (UID: \"6b353eef-b7f9-4dbc-b2aa-9860220438cc\") " pod="openshift-marketplace/community-operators-hs6pg" Dec 03 00:09:30 crc kubenswrapper[4953]: E1203 00:09:30.183756 4953 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 00:09:30.683731803 +0000 UTC m=+159.668012586 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 00:09:30 crc kubenswrapper[4953]: I1203 00:09:30.263505 4953 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-ct49l"] Dec 03 00:09:30 crc kubenswrapper[4953]: I1203 00:09:30.265032 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-ct49l" Dec 03 00:09:30 crc kubenswrapper[4953]: I1203 00:09:30.286896 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-t7ljb\" (UID: \"4313395c-8501-4b5b-a4cb-2fc986fced03\") " pod="openshift-image-registry/image-registry-697d97f7c8-t7ljb" Dec 03 00:09:30 crc kubenswrapper[4953]: I1203 00:09:30.286945 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/52fd5773-3888-49f1-813d-10d262f4c901-catalog-content\") pod \"certified-operators-xfd2l\" (UID: \"52fd5773-3888-49f1-813d-10d262f4c901\") " pod="openshift-marketplace/certified-operators-xfd2l" Dec 03 00:09:30 crc kubenswrapper[4953]: I1203 00:09:30.287002 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6b353eef-b7f9-4dbc-b2aa-9860220438cc-utilities\") pod \"community-operators-hs6pg\" (UID: \"6b353eef-b7f9-4dbc-b2aa-9860220438cc\") " pod="openshift-marketplace/community-operators-hs6pg" Dec 03 00:09:30 crc kubenswrapper[4953]: I1203 00:09:30.287028 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6b353eef-b7f9-4dbc-b2aa-9860220438cc-catalog-content\") pod \"community-operators-hs6pg\" (UID: \"6b353eef-b7f9-4dbc-b2aa-9860220438cc\") " pod="openshift-marketplace/community-operators-hs6pg" Dec 03 00:09:30 crc kubenswrapper[4953]: I1203 00:09:30.287065 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qlmtg\" (UniqueName: \"kubernetes.io/projected/52fd5773-3888-49f1-813d-10d262f4c901-kube-api-access-qlmtg\") pod \"certified-operators-xfd2l\" (UID: \"52fd5773-3888-49f1-813d-10d262f4c901\") " pod="openshift-marketplace/certified-operators-xfd2l" Dec 03 00:09:30 crc kubenswrapper[4953]: I1203 00:09:30.287085 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/52fd5773-3888-49f1-813d-10d262f4c901-utilities\") pod \"certified-operators-xfd2l\" (UID: \"52fd5773-3888-49f1-813d-10d262f4c901\") " pod="openshift-marketplace/certified-operators-xfd2l" Dec 03 00:09:30 crc kubenswrapper[4953]: I1203 00:09:30.287103 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m89m9\" (UniqueName: \"kubernetes.io/projected/6b353eef-b7f9-4dbc-b2aa-9860220438cc-kube-api-access-m89m9\") pod \"community-operators-hs6pg\" (UID: \"6b353eef-b7f9-4dbc-b2aa-9860220438cc\") " pod="openshift-marketplace/community-operators-hs6pg" Dec 03 00:09:30 crc kubenswrapper[4953]: I1203 00:09:30.287669 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6b353eef-b7f9-4dbc-b2aa-9860220438cc-utilities\") pod \"community-operators-hs6pg\" (UID: \"6b353eef-b7f9-4dbc-b2aa-9860220438cc\") " pod="openshift-marketplace/community-operators-hs6pg" Dec 03 00:09:30 crc kubenswrapper[4953]: I1203 00:09:30.287818 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6b353eef-b7f9-4dbc-b2aa-9860220438cc-catalog-content\") pod \"community-operators-hs6pg\" (UID: \"6b353eef-b7f9-4dbc-b2aa-9860220438cc\") " pod="openshift-marketplace/community-operators-hs6pg" Dec 03 00:09:30 crc kubenswrapper[4953]: E1203 00:09:30.288094 4953 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 00:09:30.788075489 +0000 UTC m=+159.772356272 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-t7ljb" (UID: "4313395c-8501-4b5b-a4cb-2fc986fced03") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 00:09:30 crc kubenswrapper[4953]: I1203 00:09:30.313420 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/52fd5773-3888-49f1-813d-10d262f4c901-utilities\") pod \"certified-operators-xfd2l\" (UID: \"52fd5773-3888-49f1-813d-10d262f4c901\") " pod="openshift-marketplace/certified-operators-xfd2l" Dec 03 00:09:30 crc kubenswrapper[4953]: I1203 00:09:30.314359 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/52fd5773-3888-49f1-813d-10d262f4c901-catalog-content\") pod \"certified-operators-xfd2l\" (UID: \"52fd5773-3888-49f1-813d-10d262f4c901\") " pod="openshift-marketplace/certified-operators-xfd2l" Dec 03 00:09:30 crc kubenswrapper[4953]: I1203 00:09:30.317113 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m89m9\" (UniqueName: \"kubernetes.io/projected/6b353eef-b7f9-4dbc-b2aa-9860220438cc-kube-api-access-m89m9\") pod \"community-operators-hs6pg\" (UID: \"6b353eef-b7f9-4dbc-b2aa-9860220438cc\") " pod="openshift-marketplace/community-operators-hs6pg" Dec 03 00:09:30 crc kubenswrapper[4953]: I1203 00:09:30.318017 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qlmtg\" (UniqueName: \"kubernetes.io/projected/52fd5773-3888-49f1-813d-10d262f4c901-kube-api-access-qlmtg\") pod \"certified-operators-xfd2l\" (UID: \"52fd5773-3888-49f1-813d-10d262f4c901\") " pod="openshift-marketplace/certified-operators-xfd2l" Dec 03 00:09:30 crc kubenswrapper[4953]: I1203 00:09:30.337183 4953 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-ct49l"] Dec 03 00:09:30 crc kubenswrapper[4953]: I1203 00:09:30.373328 4953 patch_prober.go:28] interesting pod/router-default-5444994796-r8l2l container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 03 00:09:30 crc kubenswrapper[4953]: [-]has-synced failed: reason withheld Dec 03 00:09:30 crc kubenswrapper[4953]: [+]process-running ok Dec 03 00:09:30 crc kubenswrapper[4953]: healthz check failed Dec 03 00:09:30 crc kubenswrapper[4953]: I1203 00:09:30.373402 4953 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-r8l2l" podUID="4edc57f1-5c0c-4108-bb3c-b2f104c6e08b" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 03 00:09:30 crc kubenswrapper[4953]: I1203 00:09:30.388773 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 00:09:30 crc kubenswrapper[4953]: I1203 00:09:30.389183 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3e5a99ac-9abf-4291-b773-fda0c4364959-utilities\") pod \"community-operators-ct49l\" (UID: \"3e5a99ac-9abf-4291-b773-fda0c4364959\") " pod="openshift-marketplace/community-operators-ct49l" Dec 03 00:09:30 crc kubenswrapper[4953]: I1203 00:09:30.389211 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lffhw\" (UniqueName: \"kubernetes.io/projected/3e5a99ac-9abf-4291-b773-fda0c4364959-kube-api-access-lffhw\") pod \"community-operators-ct49l\" (UID: \"3e5a99ac-9abf-4291-b773-fda0c4364959\") " pod="openshift-marketplace/community-operators-ct49l" Dec 03 00:09:30 crc kubenswrapper[4953]: I1203 00:09:30.389232 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3e5a99ac-9abf-4291-b773-fda0c4364959-catalog-content\") pod \"community-operators-ct49l\" (UID: \"3e5a99ac-9abf-4291-b773-fda0c4364959\") " pod="openshift-marketplace/community-operators-ct49l" Dec 03 00:09:30 crc kubenswrapper[4953]: E1203 00:09:30.389362 4953 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 00:09:30.889342421 +0000 UTC m=+159.873623204 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 00:09:30 crc kubenswrapper[4953]: I1203 00:09:30.442938 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-hs6pg" Dec 03 00:09:30 crc kubenswrapper[4953]: I1203 00:09:30.466540 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-xfd2l" Dec 03 00:09:30 crc kubenswrapper[4953]: I1203 00:09:30.525739 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-t7ljb\" (UID: \"4313395c-8501-4b5b-a4cb-2fc986fced03\") " pod="openshift-image-registry/image-registry-697d97f7c8-t7ljb" Dec 03 00:09:30 crc kubenswrapper[4953]: I1203 00:09:30.525777 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3e5a99ac-9abf-4291-b773-fda0c4364959-utilities\") pod \"community-operators-ct49l\" (UID: \"3e5a99ac-9abf-4291-b773-fda0c4364959\") " pod="openshift-marketplace/community-operators-ct49l" Dec 03 00:09:30 crc kubenswrapper[4953]: I1203 00:09:30.525796 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lffhw\" (UniqueName: \"kubernetes.io/projected/3e5a99ac-9abf-4291-b773-fda0c4364959-kube-api-access-lffhw\") pod \"community-operators-ct49l\" (UID: \"3e5a99ac-9abf-4291-b773-fda0c4364959\") " pod="openshift-marketplace/community-operators-ct49l" Dec 03 00:09:30 crc kubenswrapper[4953]: I1203 00:09:30.525814 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3e5a99ac-9abf-4291-b773-fda0c4364959-catalog-content\") pod \"community-operators-ct49l\" (UID: \"3e5a99ac-9abf-4291-b773-fda0c4364959\") " pod="openshift-marketplace/community-operators-ct49l" Dec 03 00:09:30 crc kubenswrapper[4953]: I1203 00:09:30.525660 4953 reconciler.go:161] "OperationExecutor.RegisterPlugin started" plugin={"SocketPath":"/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock","Timestamp":"2025-12-03T00:09:29.87840394Z","Handler":null,"Name":""} Dec 03 00:09:30 crc kubenswrapper[4953]: E1203 00:09:30.526138 4953 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 00:09:31.026126783 +0000 UTC m=+160.010407566 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-t7ljb" (UID: "4313395c-8501-4b5b-a4cb-2fc986fced03") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 00:09:30 crc kubenswrapper[4953]: I1203 00:09:30.526627 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3e5a99ac-9abf-4291-b773-fda0c4364959-utilities\") pod \"community-operators-ct49l\" (UID: \"3e5a99ac-9abf-4291-b773-fda0c4364959\") " pod="openshift-marketplace/community-operators-ct49l" Dec 03 00:09:30 crc kubenswrapper[4953]: I1203 00:09:30.526996 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3e5a99ac-9abf-4291-b773-fda0c4364959-catalog-content\") pod \"community-operators-ct49l\" (UID: \"3e5a99ac-9abf-4291-b773-fda0c4364959\") " pod="openshift-marketplace/community-operators-ct49l" Dec 03 00:09:30 crc kubenswrapper[4953]: I1203 00:09:30.528147 4953 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-js4dj"] Dec 03 00:09:30 crc kubenswrapper[4953]: I1203 00:09:30.529084 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-js4dj" Dec 03 00:09:30 crc kubenswrapper[4953]: I1203 00:09:30.546274 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lffhw\" (UniqueName: \"kubernetes.io/projected/3e5a99ac-9abf-4291-b773-fda0c4364959-kube-api-access-lffhw\") pod \"community-operators-ct49l\" (UID: \"3e5a99ac-9abf-4291-b773-fda0c4364959\") " pod="openshift-marketplace/community-operators-ct49l" Dec 03 00:09:30 crc kubenswrapper[4953]: I1203 00:09:30.558388 4953 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-js4dj"] Dec 03 00:09:30 crc kubenswrapper[4953]: I1203 00:09:30.562406 4953 csi_plugin.go:100] kubernetes.io/csi: Trying to validate a new CSI Driver with name: kubevirt.io.hostpath-provisioner endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock versions: 1.0.0 Dec 03 00:09:30 crc kubenswrapper[4953]: I1203 00:09:30.562462 4953 csi_plugin.go:113] kubernetes.io/csi: Register new plugin with name: kubevirt.io.hostpath-provisioner at endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock Dec 03 00:09:30 crc kubenswrapper[4953]: I1203 00:09:30.604736 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-ct49l" Dec 03 00:09:30 crc kubenswrapper[4953]: I1203 00:09:30.631183 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 00:09:30 crc kubenswrapper[4953]: I1203 00:09:30.631580 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f17dcaea-cb22-4359-9480-e1d129cd309e-utilities\") pod \"certified-operators-js4dj\" (UID: \"f17dcaea-cb22-4359-9480-e1d129cd309e\") " pod="openshift-marketplace/certified-operators-js4dj" Dec 03 00:09:30 crc kubenswrapper[4953]: I1203 00:09:30.631673 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f17dcaea-cb22-4359-9480-e1d129cd309e-catalog-content\") pod \"certified-operators-js4dj\" (UID: \"f17dcaea-cb22-4359-9480-e1d129cd309e\") " pod="openshift-marketplace/certified-operators-js4dj" Dec 03 00:09:30 crc kubenswrapper[4953]: I1203 00:09:30.631721 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jfmxh\" (UniqueName: \"kubernetes.io/projected/f17dcaea-cb22-4359-9480-e1d129cd309e-kube-api-access-jfmxh\") pod \"certified-operators-js4dj\" (UID: \"f17dcaea-cb22-4359-9480-e1d129cd309e\") " pod="openshift-marketplace/certified-operators-js4dj" Dec 03 00:09:30 crc kubenswrapper[4953]: I1203 00:09:30.649878 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Dec 03 00:09:30 crc kubenswrapper[4953]: I1203 00:09:30.670444 4953 generic.go:334] "Generic (PLEG): container finished" podID="7b91526b-2492-4c17-949b-0ba3f8c221ea" containerID="c9bb8e9e994fc6c10d6aab1c55ae333228332a4ed3369558948c3249ba7c71af" exitCode=0 Dec 03 00:09:30 crc kubenswrapper[4953]: I1203 00:09:30.670503 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29412000-mqgxv" event={"ID":"7b91526b-2492-4c17-949b-0ba3f8c221ea","Type":"ContainerDied","Data":"c9bb8e9e994fc6c10d6aab1c55ae333228332a4ed3369558948c3249ba7c71af"} Dec 03 00:09:30 crc kubenswrapper[4953]: I1203 00:09:30.674911 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-2xt2r" event={"ID":"73033ee8-cae5-4acc-8c77-60905eed62f2","Type":"ContainerStarted","Data":"039b28ced236cf37cd66e9761fcbaa5dc44999d8a6938b928f32751d732fc01f"} Dec 03 00:09:30 crc kubenswrapper[4953]: I1203 00:09:30.733780 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f17dcaea-cb22-4359-9480-e1d129cd309e-catalog-content\") pod \"certified-operators-js4dj\" (UID: \"f17dcaea-cb22-4359-9480-e1d129cd309e\") " pod="openshift-marketplace/certified-operators-js4dj" Dec 03 00:09:30 crc kubenswrapper[4953]: I1203 00:09:30.733863 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jfmxh\" (UniqueName: \"kubernetes.io/projected/f17dcaea-cb22-4359-9480-e1d129cd309e-kube-api-access-jfmxh\") pod \"certified-operators-js4dj\" (UID: \"f17dcaea-cb22-4359-9480-e1d129cd309e\") " pod="openshift-marketplace/certified-operators-js4dj" Dec 03 00:09:30 crc kubenswrapper[4953]: I1203 00:09:30.733893 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-t7ljb\" (UID: \"4313395c-8501-4b5b-a4cb-2fc986fced03\") " pod="openshift-image-registry/image-registry-697d97f7c8-t7ljb" Dec 03 00:09:30 crc kubenswrapper[4953]: I1203 00:09:30.733931 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f17dcaea-cb22-4359-9480-e1d129cd309e-utilities\") pod \"certified-operators-js4dj\" (UID: \"f17dcaea-cb22-4359-9480-e1d129cd309e\") " pod="openshift-marketplace/certified-operators-js4dj" Dec 03 00:09:30 crc kubenswrapper[4953]: I1203 00:09:30.734617 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f17dcaea-cb22-4359-9480-e1d129cd309e-utilities\") pod \"certified-operators-js4dj\" (UID: \"f17dcaea-cb22-4359-9480-e1d129cd309e\") " pod="openshift-marketplace/certified-operators-js4dj" Dec 03 00:09:30 crc kubenswrapper[4953]: I1203 00:09:30.735211 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f17dcaea-cb22-4359-9480-e1d129cd309e-catalog-content\") pod \"certified-operators-js4dj\" (UID: \"f17dcaea-cb22-4359-9480-e1d129cd309e\") " pod="openshift-marketplace/certified-operators-js4dj" Dec 03 00:09:30 crc kubenswrapper[4953]: I1203 00:09:30.816785 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jfmxh\" (UniqueName: \"kubernetes.io/projected/f17dcaea-cb22-4359-9480-e1d129cd309e-kube-api-access-jfmxh\") pod \"certified-operators-js4dj\" (UID: \"f17dcaea-cb22-4359-9480-e1d129cd309e\") " pod="openshift-marketplace/certified-operators-js4dj" Dec 03 00:09:30 crc kubenswrapper[4953]: I1203 00:09:30.904494 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-js4dj" Dec 03 00:09:31 crc kubenswrapper[4953]: I1203 00:09:31.050870 4953 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 03 00:09:31 crc kubenswrapper[4953]: I1203 00:09:31.051194 4953 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-t7ljb\" (UID: \"4313395c-8501-4b5b-a4cb-2fc986fced03\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount\"" pod="openshift-image-registry/image-registry-697d97f7c8-t7ljb" Dec 03 00:09:31 crc kubenswrapper[4953]: I1203 00:09:31.150549 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-t7ljb\" (UID: \"4313395c-8501-4b5b-a4cb-2fc986fced03\") " pod="openshift-image-registry/image-registry-697d97f7c8-t7ljb" Dec 03 00:09:31 crc kubenswrapper[4953]: I1203 00:09:31.184163 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-t7ljb" Dec 03 00:09:31 crc kubenswrapper[4953]: I1203 00:09:31.235028 4953 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-config-operator/openshift-config-operator-7777fb866f-l7m9z" Dec 03 00:09:31 crc kubenswrapper[4953]: I1203 00:09:31.360167 4953 patch_prober.go:28] interesting pod/router-default-5444994796-r8l2l container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 03 00:09:31 crc kubenswrapper[4953]: [-]has-synced failed: reason withheld Dec 03 00:09:31 crc kubenswrapper[4953]: [+]process-running ok Dec 03 00:09:31 crc kubenswrapper[4953]: healthz check failed Dec 03 00:09:31 crc kubenswrapper[4953]: I1203 00:09:31.360792 4953 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-r8l2l" podUID="4edc57f1-5c0c-4108-bb3c-b2f104c6e08b" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 03 00:09:31 crc kubenswrapper[4953]: I1203 00:09:31.370835 4953 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f668bae-612b-4b75-9490-919e737c6a3b" path="/var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes" Dec 03 00:09:31 crc kubenswrapper[4953]: I1203 00:09:31.418824 4953 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-xfd2l"] Dec 03 00:09:31 crc kubenswrapper[4953]: I1203 00:09:31.430390 4953 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-ct49l"] Dec 03 00:09:31 crc kubenswrapper[4953]: I1203 00:09:31.495662 4953 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-hs6pg"] Dec 03 00:09:31 crc kubenswrapper[4953]: I1203 00:09:31.634902 4953 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-t7ljb"] Dec 03 00:09:31 crc kubenswrapper[4953]: I1203 00:09:31.680756 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xfd2l" event={"ID":"52fd5773-3888-49f1-813d-10d262f4c901","Type":"ContainerStarted","Data":"37d8cbc201645bf6b66467acb62ab2437ecd3b0b9d98aa4cfefee38db9f2b0ee"} Dec 03 00:09:31 crc kubenswrapper[4953]: I1203 00:09:31.683816 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ct49l" event={"ID":"3e5a99ac-9abf-4291-b773-fda0c4364959","Type":"ContainerStarted","Data":"78ef4c2de9211ad620e2f2ce4fa3886bea7baa989225de5e229e15ace101e0b7"} Dec 03 00:09:31 crc kubenswrapper[4953]: I1203 00:09:31.685513 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hs6pg" event={"ID":"6b353eef-b7f9-4dbc-b2aa-9860220438cc","Type":"ContainerStarted","Data":"224b30baa9aff1a09e27d6c314399ce59aee5ea2b5e60b153e20cbadcb677515"} Dec 03 00:09:31 crc kubenswrapper[4953]: W1203 00:09:31.760083 4953 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4313395c_8501_4b5b_a4cb_2fc986fced03.slice/crio-61aa05ede5ab5d75673b721f1a77b40fa41eb1c55d50112cb6687d7cfa0d7305 WatchSource:0}: Error finding container 61aa05ede5ab5d75673b721f1a77b40fa41eb1c55d50112cb6687d7cfa0d7305: Status 404 returned error can't find the container with id 61aa05ede5ab5d75673b721f1a77b40fa41eb1c55d50112cb6687d7cfa0d7305 Dec 03 00:09:31 crc kubenswrapper[4953]: I1203 00:09:31.766348 4953 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-js4dj"] Dec 03 00:09:31 crc kubenswrapper[4953]: I1203 00:09:31.884958 4953 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-5zxsj"] Dec 03 00:09:31 crc kubenswrapper[4953]: I1203 00:09:31.886533 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-5zxsj" Dec 03 00:09:31 crc kubenswrapper[4953]: I1203 00:09:31.890057 4953 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Dec 03 00:09:31 crc kubenswrapper[4953]: I1203 00:09:31.900690 4953 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-5zxsj"] Dec 03 00:09:31 crc kubenswrapper[4953]: I1203 00:09:31.915693 4953 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Dec 03 00:09:31 crc kubenswrapper[4953]: I1203 00:09:31.916388 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 03 00:09:31 crc kubenswrapper[4953]: I1203 00:09:31.922752 4953 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Dec 03 00:09:31 crc kubenswrapper[4953]: I1203 00:09:31.923315 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager"/"kube-root-ca.crt" Dec 03 00:09:31 crc kubenswrapper[4953]: I1203 00:09:31.923670 4953 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager"/"installer-sa-dockercfg-kjl2n" Dec 03 00:09:31 crc kubenswrapper[4953]: I1203 00:09:31.976033 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/30fac6bb-4dfc-44c7-b8af-0342a8d504b7-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"30fac6bb-4dfc-44c7-b8af-0342a8d504b7\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 03 00:09:31 crc kubenswrapper[4953]: I1203 00:09:31.976297 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/30fac6bb-4dfc-44c7-b8af-0342a8d504b7-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"30fac6bb-4dfc-44c7-b8af-0342a8d504b7\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 03 00:09:31 crc kubenswrapper[4953]: I1203 00:09:31.976390 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kv2jf\" (UniqueName: \"kubernetes.io/projected/e98b8740-f6fc-4db0-a807-664cf085391c-kube-api-access-kv2jf\") pod \"redhat-marketplace-5zxsj\" (UID: \"e98b8740-f6fc-4db0-a807-664cf085391c\") " pod="openshift-marketplace/redhat-marketplace-5zxsj" Dec 03 00:09:31 crc kubenswrapper[4953]: I1203 00:09:31.976474 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e98b8740-f6fc-4db0-a807-664cf085391c-catalog-content\") pod \"redhat-marketplace-5zxsj\" (UID: \"e98b8740-f6fc-4db0-a807-664cf085391c\") " pod="openshift-marketplace/redhat-marketplace-5zxsj" Dec 03 00:09:31 crc kubenswrapper[4953]: I1203 00:09:31.976576 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e98b8740-f6fc-4db0-a807-664cf085391c-utilities\") pod \"redhat-marketplace-5zxsj\" (UID: \"e98b8740-f6fc-4db0-a807-664cf085391c\") " pod="openshift-marketplace/redhat-marketplace-5zxsj" Dec 03 00:09:32 crc kubenswrapper[4953]: I1203 00:09:32.077481 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e98b8740-f6fc-4db0-a807-664cf085391c-utilities\") pod \"redhat-marketplace-5zxsj\" (UID: \"e98b8740-f6fc-4db0-a807-664cf085391c\") " pod="openshift-marketplace/redhat-marketplace-5zxsj" Dec 03 00:09:32 crc kubenswrapper[4953]: I1203 00:09:32.077921 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/30fac6bb-4dfc-44c7-b8af-0342a8d504b7-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"30fac6bb-4dfc-44c7-b8af-0342a8d504b7\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 03 00:09:32 crc kubenswrapper[4953]: I1203 00:09:32.077941 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/30fac6bb-4dfc-44c7-b8af-0342a8d504b7-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"30fac6bb-4dfc-44c7-b8af-0342a8d504b7\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 03 00:09:32 crc kubenswrapper[4953]: I1203 00:09:32.077975 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kv2jf\" (UniqueName: \"kubernetes.io/projected/e98b8740-f6fc-4db0-a807-664cf085391c-kube-api-access-kv2jf\") pod \"redhat-marketplace-5zxsj\" (UID: \"e98b8740-f6fc-4db0-a807-664cf085391c\") " pod="openshift-marketplace/redhat-marketplace-5zxsj" Dec 03 00:09:32 crc kubenswrapper[4953]: I1203 00:09:32.078017 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e98b8740-f6fc-4db0-a807-664cf085391c-catalog-content\") pod \"redhat-marketplace-5zxsj\" (UID: \"e98b8740-f6fc-4db0-a807-664cf085391c\") " pod="openshift-marketplace/redhat-marketplace-5zxsj" Dec 03 00:09:32 crc kubenswrapper[4953]: I1203 00:09:32.078025 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/30fac6bb-4dfc-44c7-b8af-0342a8d504b7-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"30fac6bb-4dfc-44c7-b8af-0342a8d504b7\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 03 00:09:32 crc kubenswrapper[4953]: I1203 00:09:32.078422 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e98b8740-f6fc-4db0-a807-664cf085391c-utilities\") pod \"redhat-marketplace-5zxsj\" (UID: \"e98b8740-f6fc-4db0-a807-664cf085391c\") " pod="openshift-marketplace/redhat-marketplace-5zxsj" Dec 03 00:09:32 crc kubenswrapper[4953]: I1203 00:09:32.079994 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e98b8740-f6fc-4db0-a807-664cf085391c-catalog-content\") pod \"redhat-marketplace-5zxsj\" (UID: \"e98b8740-f6fc-4db0-a807-664cf085391c\") " pod="openshift-marketplace/redhat-marketplace-5zxsj" Dec 03 00:09:32 crc kubenswrapper[4953]: I1203 00:09:32.107441 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kv2jf\" (UniqueName: \"kubernetes.io/projected/e98b8740-f6fc-4db0-a807-664cf085391c-kube-api-access-kv2jf\") pod \"redhat-marketplace-5zxsj\" (UID: \"e98b8740-f6fc-4db0-a807-664cf085391c\") " pod="openshift-marketplace/redhat-marketplace-5zxsj" Dec 03 00:09:32 crc kubenswrapper[4953]: I1203 00:09:32.107698 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/30fac6bb-4dfc-44c7-b8af-0342a8d504b7-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"30fac6bb-4dfc-44c7-b8af-0342a8d504b7\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 03 00:09:32 crc kubenswrapper[4953]: I1203 00:09:32.159567 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-5zxsj" Dec 03 00:09:32 crc kubenswrapper[4953]: I1203 00:09:32.231160 4953 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29412000-mqgxv" Dec 03 00:09:32 crc kubenswrapper[4953]: I1203 00:09:32.234715 4953 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-6kz7m"] Dec 03 00:09:32 crc kubenswrapper[4953]: E1203 00:09:32.234940 4953 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7b91526b-2492-4c17-949b-0ba3f8c221ea" containerName="collect-profiles" Dec 03 00:09:32 crc kubenswrapper[4953]: I1203 00:09:32.234953 4953 state_mem.go:107] "Deleted CPUSet assignment" podUID="7b91526b-2492-4c17-949b-0ba3f8c221ea" containerName="collect-profiles" Dec 03 00:09:32 crc kubenswrapper[4953]: I1203 00:09:32.236633 4953 memory_manager.go:354] "RemoveStaleState removing state" podUID="7b91526b-2492-4c17-949b-0ba3f8c221ea" containerName="collect-profiles" Dec 03 00:09:32 crc kubenswrapper[4953]: I1203 00:09:32.237933 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-6kz7m" Dec 03 00:09:32 crc kubenswrapper[4953]: I1203 00:09:32.239512 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 03 00:09:32 crc kubenswrapper[4953]: I1203 00:09:32.297575 4953 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-6kz7m"] Dec 03 00:09:32 crc kubenswrapper[4953]: I1203 00:09:32.350994 4953 patch_prober.go:28] interesting pod/router-default-5444994796-r8l2l container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 03 00:09:32 crc kubenswrapper[4953]: [-]has-synced failed: reason withheld Dec 03 00:09:32 crc kubenswrapper[4953]: [+]process-running ok Dec 03 00:09:32 crc kubenswrapper[4953]: healthz check failed Dec 03 00:09:32 crc kubenswrapper[4953]: I1203 00:09:32.351058 4953 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-r8l2l" podUID="4edc57f1-5c0c-4108-bb3c-b2f104c6e08b" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 03 00:09:32 crc kubenswrapper[4953]: I1203 00:09:32.394439 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6j5fc\" (UniqueName: \"kubernetes.io/projected/7b91526b-2492-4c17-949b-0ba3f8c221ea-kube-api-access-6j5fc\") pod \"7b91526b-2492-4c17-949b-0ba3f8c221ea\" (UID: \"7b91526b-2492-4c17-949b-0ba3f8c221ea\") " Dec 03 00:09:32 crc kubenswrapper[4953]: I1203 00:09:32.395831 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/7b91526b-2492-4c17-949b-0ba3f8c221ea-secret-volume\") pod \"7b91526b-2492-4c17-949b-0ba3f8c221ea\" (UID: \"7b91526b-2492-4c17-949b-0ba3f8c221ea\") " Dec 03 00:09:32 crc kubenswrapper[4953]: I1203 00:09:32.395902 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/7b91526b-2492-4c17-949b-0ba3f8c221ea-config-volume\") pod \"7b91526b-2492-4c17-949b-0ba3f8c221ea\" (UID: \"7b91526b-2492-4c17-949b-0ba3f8c221ea\") " Dec 03 00:09:32 crc kubenswrapper[4953]: I1203 00:09:32.396038 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/496add53-6511-4560-82c5-3efeb65935f0-utilities\") pod \"redhat-marketplace-6kz7m\" (UID: \"496add53-6511-4560-82c5-3efeb65935f0\") " pod="openshift-marketplace/redhat-marketplace-6kz7m" Dec 03 00:09:32 crc kubenswrapper[4953]: I1203 00:09:32.396085 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/496add53-6511-4560-82c5-3efeb65935f0-catalog-content\") pod \"redhat-marketplace-6kz7m\" (UID: \"496add53-6511-4560-82c5-3efeb65935f0\") " pod="openshift-marketplace/redhat-marketplace-6kz7m" Dec 03 00:09:32 crc kubenswrapper[4953]: I1203 00:09:32.396136 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gjvz8\" (UniqueName: \"kubernetes.io/projected/496add53-6511-4560-82c5-3efeb65935f0-kube-api-access-gjvz8\") pod \"redhat-marketplace-6kz7m\" (UID: \"496add53-6511-4560-82c5-3efeb65935f0\") " pod="openshift-marketplace/redhat-marketplace-6kz7m" Dec 03 00:09:32 crc kubenswrapper[4953]: I1203 00:09:32.398809 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7b91526b-2492-4c17-949b-0ba3f8c221ea-config-volume" (OuterVolumeSpecName: "config-volume") pod "7b91526b-2492-4c17-949b-0ba3f8c221ea" (UID: "7b91526b-2492-4c17-949b-0ba3f8c221ea"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:09:32 crc kubenswrapper[4953]: I1203 00:09:32.399729 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7b91526b-2492-4c17-949b-0ba3f8c221ea-kube-api-access-6j5fc" (OuterVolumeSpecName: "kube-api-access-6j5fc") pod "7b91526b-2492-4c17-949b-0ba3f8c221ea" (UID: "7b91526b-2492-4c17-949b-0ba3f8c221ea"). InnerVolumeSpecName "kube-api-access-6j5fc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:09:32 crc kubenswrapper[4953]: I1203 00:09:32.414123 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7b91526b-2492-4c17-949b-0ba3f8c221ea-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "7b91526b-2492-4c17-949b-0ba3f8c221ea" (UID: "7b91526b-2492-4c17-949b-0ba3f8c221ea"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:09:32 crc kubenswrapper[4953]: I1203 00:09:32.415488 4953 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-5zxsj"] Dec 03 00:09:32 crc kubenswrapper[4953]: I1203 00:09:32.497238 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/496add53-6511-4560-82c5-3efeb65935f0-utilities\") pod \"redhat-marketplace-6kz7m\" (UID: \"496add53-6511-4560-82c5-3efeb65935f0\") " pod="openshift-marketplace/redhat-marketplace-6kz7m" Dec 03 00:09:32 crc kubenswrapper[4953]: I1203 00:09:32.497306 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/496add53-6511-4560-82c5-3efeb65935f0-catalog-content\") pod \"redhat-marketplace-6kz7m\" (UID: \"496add53-6511-4560-82c5-3efeb65935f0\") " pod="openshift-marketplace/redhat-marketplace-6kz7m" Dec 03 00:09:32 crc kubenswrapper[4953]: I1203 00:09:32.497336 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gjvz8\" (UniqueName: \"kubernetes.io/projected/496add53-6511-4560-82c5-3efeb65935f0-kube-api-access-gjvz8\") pod \"redhat-marketplace-6kz7m\" (UID: \"496add53-6511-4560-82c5-3efeb65935f0\") " pod="openshift-marketplace/redhat-marketplace-6kz7m" Dec 03 00:09:32 crc kubenswrapper[4953]: I1203 00:09:32.497426 4953 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6j5fc\" (UniqueName: \"kubernetes.io/projected/7b91526b-2492-4c17-949b-0ba3f8c221ea-kube-api-access-6j5fc\") on node \"crc\" DevicePath \"\"" Dec 03 00:09:32 crc kubenswrapper[4953]: I1203 00:09:32.497442 4953 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/7b91526b-2492-4c17-949b-0ba3f8c221ea-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 03 00:09:32 crc kubenswrapper[4953]: I1203 00:09:32.497454 4953 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/7b91526b-2492-4c17-949b-0ba3f8c221ea-config-volume\") on node \"crc\" DevicePath \"\"" Dec 03 00:09:32 crc kubenswrapper[4953]: I1203 00:09:32.497891 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/496add53-6511-4560-82c5-3efeb65935f0-utilities\") pod \"redhat-marketplace-6kz7m\" (UID: \"496add53-6511-4560-82c5-3efeb65935f0\") " pod="openshift-marketplace/redhat-marketplace-6kz7m" Dec 03 00:09:32 crc kubenswrapper[4953]: I1203 00:09:32.498411 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/496add53-6511-4560-82c5-3efeb65935f0-catalog-content\") pod \"redhat-marketplace-6kz7m\" (UID: \"496add53-6511-4560-82c5-3efeb65935f0\") " pod="openshift-marketplace/redhat-marketplace-6kz7m" Dec 03 00:09:32 crc kubenswrapper[4953]: I1203 00:09:32.515613 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gjvz8\" (UniqueName: \"kubernetes.io/projected/496add53-6511-4560-82c5-3efeb65935f0-kube-api-access-gjvz8\") pod \"redhat-marketplace-6kz7m\" (UID: \"496add53-6511-4560-82c5-3efeb65935f0\") " pod="openshift-marketplace/redhat-marketplace-6kz7m" Dec 03 00:09:32 crc kubenswrapper[4953]: I1203 00:09:32.552764 4953 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Dec 03 00:09:32 crc kubenswrapper[4953]: W1203 00:09:32.560445 4953 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-pod30fac6bb_4dfc_44c7_b8af_0342a8d504b7.slice/crio-c6f1decaff066a9700bed15273ab687fcdc3ae7bb81fd359a8ba51e570df631a WatchSource:0}: Error finding container c6f1decaff066a9700bed15273ab687fcdc3ae7bb81fd359a8ba51e570df631a: Status 404 returned error can't find the container with id c6f1decaff066a9700bed15273ab687fcdc3ae7bb81fd359a8ba51e570df631a Dec 03 00:09:32 crc kubenswrapper[4953]: I1203 00:09:32.631565 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-6kz7m" Dec 03 00:09:32 crc kubenswrapper[4953]: I1203 00:09:32.699344 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-t7ljb" event={"ID":"4313395c-8501-4b5b-a4cb-2fc986fced03","Type":"ContainerStarted","Data":"61aa05ede5ab5d75673b721f1a77b40fa41eb1c55d50112cb6687d7cfa0d7305"} Dec 03 00:09:32 crc kubenswrapper[4953]: I1203 00:09:32.702775 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5zxsj" event={"ID":"e98b8740-f6fc-4db0-a807-664cf085391c","Type":"ContainerStarted","Data":"c747d83b368b5a45d4daead722975738b091c4d0e85f6a787fc22c844d1bd7dd"} Dec 03 00:09:32 crc kubenswrapper[4953]: I1203 00:09:32.704042 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"30fac6bb-4dfc-44c7-b8af-0342a8d504b7","Type":"ContainerStarted","Data":"c6f1decaff066a9700bed15273ab687fcdc3ae7bb81fd359a8ba51e570df631a"} Dec 03 00:09:32 crc kubenswrapper[4953]: I1203 00:09:32.706095 4953 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29412000-mqgxv" Dec 03 00:09:32 crc kubenswrapper[4953]: I1203 00:09:32.707135 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29412000-mqgxv" event={"ID":"7b91526b-2492-4c17-949b-0ba3f8c221ea","Type":"ContainerDied","Data":"cb0ad0b7aefd5f2be56b688771208a14b485d7dfd5b9b676113f11e5b27a7302"} Dec 03 00:09:32 crc kubenswrapper[4953]: I1203 00:09:32.707195 4953 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cb0ad0b7aefd5f2be56b688771208a14b485d7dfd5b9b676113f11e5b27a7302" Dec 03 00:09:32 crc kubenswrapper[4953]: I1203 00:09:32.708198 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-js4dj" event={"ID":"f17dcaea-cb22-4359-9480-e1d129cd309e","Type":"ContainerStarted","Data":"00378a6caa696020038368823cb33ae715cae36617efa82c26df2c5c5d2708e6"} Dec 03 00:09:33 crc kubenswrapper[4953]: I1203 00:09:33.048615 4953 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-6kz7m"] Dec 03 00:09:33 crc kubenswrapper[4953]: W1203 00:09:33.058828 4953 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod496add53_6511_4560_82c5_3efeb65935f0.slice/crio-eadf7e7493d10cc4697afbeac29011f04e4fa41b0dc506c07df8179ab4703e75 WatchSource:0}: Error finding container eadf7e7493d10cc4697afbeac29011f04e4fa41b0dc506c07df8179ab4703e75: Status 404 returned error can't find the container with id eadf7e7493d10cc4697afbeac29011f04e4fa41b0dc506c07df8179ab4703e75 Dec 03 00:09:33 crc kubenswrapper[4953]: I1203 00:09:33.289079 4953 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-dq6nn"] Dec 03 00:09:33 crc kubenswrapper[4953]: I1203 00:09:33.290539 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-dq6nn" Dec 03 00:09:33 crc kubenswrapper[4953]: I1203 00:09:33.293553 4953 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Dec 03 00:09:33 crc kubenswrapper[4953]: I1203 00:09:33.307856 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8d5tx\" (UniqueName: \"kubernetes.io/projected/9064a5ab-24ee-4258-8363-6ce4ffa54b3f-kube-api-access-8d5tx\") pod \"redhat-operators-dq6nn\" (UID: \"9064a5ab-24ee-4258-8363-6ce4ffa54b3f\") " pod="openshift-marketplace/redhat-operators-dq6nn" Dec 03 00:09:33 crc kubenswrapper[4953]: I1203 00:09:33.308090 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9064a5ab-24ee-4258-8363-6ce4ffa54b3f-utilities\") pod \"redhat-operators-dq6nn\" (UID: \"9064a5ab-24ee-4258-8363-6ce4ffa54b3f\") " pod="openshift-marketplace/redhat-operators-dq6nn" Dec 03 00:09:33 crc kubenswrapper[4953]: I1203 00:09:33.308167 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9064a5ab-24ee-4258-8363-6ce4ffa54b3f-catalog-content\") pod \"redhat-operators-dq6nn\" (UID: \"9064a5ab-24ee-4258-8363-6ce4ffa54b3f\") " pod="openshift-marketplace/redhat-operators-dq6nn" Dec 03 00:09:33 crc kubenswrapper[4953]: I1203 00:09:33.313652 4953 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-dq6nn"] Dec 03 00:09:33 crc kubenswrapper[4953]: I1203 00:09:33.351371 4953 patch_prober.go:28] interesting pod/router-default-5444994796-r8l2l container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 03 00:09:33 crc kubenswrapper[4953]: [-]has-synced failed: reason withheld Dec 03 00:09:33 crc kubenswrapper[4953]: [+]process-running ok Dec 03 00:09:33 crc kubenswrapper[4953]: healthz check failed Dec 03 00:09:33 crc kubenswrapper[4953]: I1203 00:09:33.351900 4953 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-r8l2l" podUID="4edc57f1-5c0c-4108-bb3c-b2f104c6e08b" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 03 00:09:33 crc kubenswrapper[4953]: I1203 00:09:33.410126 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8d5tx\" (UniqueName: \"kubernetes.io/projected/9064a5ab-24ee-4258-8363-6ce4ffa54b3f-kube-api-access-8d5tx\") pod \"redhat-operators-dq6nn\" (UID: \"9064a5ab-24ee-4258-8363-6ce4ffa54b3f\") " pod="openshift-marketplace/redhat-operators-dq6nn" Dec 03 00:09:33 crc kubenswrapper[4953]: I1203 00:09:33.410252 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9064a5ab-24ee-4258-8363-6ce4ffa54b3f-utilities\") pod \"redhat-operators-dq6nn\" (UID: \"9064a5ab-24ee-4258-8363-6ce4ffa54b3f\") " pod="openshift-marketplace/redhat-operators-dq6nn" Dec 03 00:09:33 crc kubenswrapper[4953]: I1203 00:09:33.410280 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9064a5ab-24ee-4258-8363-6ce4ffa54b3f-catalog-content\") pod \"redhat-operators-dq6nn\" (UID: \"9064a5ab-24ee-4258-8363-6ce4ffa54b3f\") " pod="openshift-marketplace/redhat-operators-dq6nn" Dec 03 00:09:33 crc kubenswrapper[4953]: I1203 00:09:33.411098 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9064a5ab-24ee-4258-8363-6ce4ffa54b3f-catalog-content\") pod \"redhat-operators-dq6nn\" (UID: \"9064a5ab-24ee-4258-8363-6ce4ffa54b3f\") " pod="openshift-marketplace/redhat-operators-dq6nn" Dec 03 00:09:33 crc kubenswrapper[4953]: I1203 00:09:33.411207 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9064a5ab-24ee-4258-8363-6ce4ffa54b3f-utilities\") pod \"redhat-operators-dq6nn\" (UID: \"9064a5ab-24ee-4258-8363-6ce4ffa54b3f\") " pod="openshift-marketplace/redhat-operators-dq6nn" Dec 03 00:09:33 crc kubenswrapper[4953]: I1203 00:09:33.437750 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8d5tx\" (UniqueName: \"kubernetes.io/projected/9064a5ab-24ee-4258-8363-6ce4ffa54b3f-kube-api-access-8d5tx\") pod \"redhat-operators-dq6nn\" (UID: \"9064a5ab-24ee-4258-8363-6ce4ffa54b3f\") " pod="openshift-marketplace/redhat-operators-dq6nn" Dec 03 00:09:33 crc kubenswrapper[4953]: I1203 00:09:33.495640 4953 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-apiserver/apiserver-76f77b778f-hzbcl" Dec 03 00:09:33 crc kubenswrapper[4953]: I1203 00:09:33.503459 4953 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-apiserver/apiserver-76f77b778f-hzbcl" Dec 03 00:09:33 crc kubenswrapper[4953]: I1203 00:09:33.610401 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-dq6nn" Dec 03 00:09:33 crc kubenswrapper[4953]: I1203 00:09:33.644810 4953 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-q94c6"] Dec 03 00:09:33 crc kubenswrapper[4953]: I1203 00:09:33.646092 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-q94c6" Dec 03 00:09:33 crc kubenswrapper[4953]: I1203 00:09:33.678079 4953 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-q94c6"] Dec 03 00:09:33 crc kubenswrapper[4953]: I1203 00:09:33.758015 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ct49l" event={"ID":"3e5a99ac-9abf-4291-b773-fda0c4364959","Type":"ContainerStarted","Data":"9fbad7d01aef12b68dbe8856ab62492197d7d0237185915678fa26b4609f5e54"} Dec 03 00:09:33 crc kubenswrapper[4953]: I1203 00:09:33.759181 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6kz7m" event={"ID":"496add53-6511-4560-82c5-3efeb65935f0","Type":"ContainerStarted","Data":"eadf7e7493d10cc4697afbeac29011f04e4fa41b0dc506c07df8179ab4703e75"} Dec 03 00:09:33 crc kubenswrapper[4953]: I1203 00:09:33.783288 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-2xt2r" event={"ID":"73033ee8-cae5-4acc-8c77-60905eed62f2","Type":"ContainerStarted","Data":"ffeedc4fb017c6f90ac7c2cc6aefb7c9442e2f364dc1679767e455d1917fbf48"} Dec 03 00:09:33 crc kubenswrapper[4953]: I1203 00:09:33.833334 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ghrbv\" (UniqueName: \"kubernetes.io/projected/be2fa7e6-200e-44f0-9b40-57ed2d30db95-kube-api-access-ghrbv\") pod \"redhat-operators-q94c6\" (UID: \"be2fa7e6-200e-44f0-9b40-57ed2d30db95\") " pod="openshift-marketplace/redhat-operators-q94c6" Dec 03 00:09:33 crc kubenswrapper[4953]: I1203 00:09:33.833436 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/be2fa7e6-200e-44f0-9b40-57ed2d30db95-utilities\") pod \"redhat-operators-q94c6\" (UID: \"be2fa7e6-200e-44f0-9b40-57ed2d30db95\") " pod="openshift-marketplace/redhat-operators-q94c6" Dec 03 00:09:33 crc kubenswrapper[4953]: I1203 00:09:33.833470 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/be2fa7e6-200e-44f0-9b40-57ed2d30db95-catalog-content\") pod \"redhat-operators-q94c6\" (UID: \"be2fa7e6-200e-44f0-9b40-57ed2d30db95\") " pod="openshift-marketplace/redhat-operators-q94c6" Dec 03 00:09:33 crc kubenswrapper[4953]: I1203 00:09:33.934406 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/be2fa7e6-200e-44f0-9b40-57ed2d30db95-catalog-content\") pod \"redhat-operators-q94c6\" (UID: \"be2fa7e6-200e-44f0-9b40-57ed2d30db95\") " pod="openshift-marketplace/redhat-operators-q94c6" Dec 03 00:09:33 crc kubenswrapper[4953]: I1203 00:09:33.935433 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/be2fa7e6-200e-44f0-9b40-57ed2d30db95-catalog-content\") pod \"redhat-operators-q94c6\" (UID: \"be2fa7e6-200e-44f0-9b40-57ed2d30db95\") " pod="openshift-marketplace/redhat-operators-q94c6" Dec 03 00:09:33 crc kubenswrapper[4953]: I1203 00:09:33.935487 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ghrbv\" (UniqueName: \"kubernetes.io/projected/be2fa7e6-200e-44f0-9b40-57ed2d30db95-kube-api-access-ghrbv\") pod \"redhat-operators-q94c6\" (UID: \"be2fa7e6-200e-44f0-9b40-57ed2d30db95\") " pod="openshift-marketplace/redhat-operators-q94c6" Dec 03 00:09:33 crc kubenswrapper[4953]: I1203 00:09:33.935585 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/be2fa7e6-200e-44f0-9b40-57ed2d30db95-utilities\") pod \"redhat-operators-q94c6\" (UID: \"be2fa7e6-200e-44f0-9b40-57ed2d30db95\") " pod="openshift-marketplace/redhat-operators-q94c6" Dec 03 00:09:33 crc kubenswrapper[4953]: I1203 00:09:33.936775 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/be2fa7e6-200e-44f0-9b40-57ed2d30db95-utilities\") pod \"redhat-operators-q94c6\" (UID: \"be2fa7e6-200e-44f0-9b40-57ed2d30db95\") " pod="openshift-marketplace/redhat-operators-q94c6" Dec 03 00:09:33 crc kubenswrapper[4953]: I1203 00:09:33.959863 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ghrbv\" (UniqueName: \"kubernetes.io/projected/be2fa7e6-200e-44f0-9b40-57ed2d30db95-kube-api-access-ghrbv\") pod \"redhat-operators-q94c6\" (UID: \"be2fa7e6-200e-44f0-9b40-57ed2d30db95\") " pod="openshift-marketplace/redhat-operators-q94c6" Dec 03 00:09:33 crc kubenswrapper[4953]: I1203 00:09:33.975184 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-q94c6" Dec 03 00:09:34 crc kubenswrapper[4953]: I1203 00:09:34.063119 4953 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-dq6nn"] Dec 03 00:09:34 crc kubenswrapper[4953]: I1203 00:09:34.486291 4953 patch_prober.go:28] interesting pod/router-default-5444994796-r8l2l container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 03 00:09:34 crc kubenswrapper[4953]: [-]has-synced failed: reason withheld Dec 03 00:09:34 crc kubenswrapper[4953]: [+]process-running ok Dec 03 00:09:34 crc kubenswrapper[4953]: healthz check failed Dec 03 00:09:34 crc kubenswrapper[4953]: I1203 00:09:34.486929 4953 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-r8l2l" podUID="4edc57f1-5c0c-4108-bb3c-b2f104c6e08b" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 03 00:09:34 crc kubenswrapper[4953]: I1203 00:09:34.572385 4953 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-q94c6"] Dec 03 00:09:34 crc kubenswrapper[4953]: I1203 00:09:34.790862 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-t7ljb" event={"ID":"4313395c-8501-4b5b-a4cb-2fc986fced03","Type":"ContainerStarted","Data":"854d5d14e7ff3b9f049285ae27971f771bd934c5e0af219c59d73e018887a492"} Dec 03 00:09:34 crc kubenswrapper[4953]: W1203 00:09:34.795259 4953 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podbe2fa7e6_200e_44f0_9b40_57ed2d30db95.slice/crio-c64ff0759b4e0582da669df0d162803598a061dd7c8baa5aef06f79aba6470ac WatchSource:0}: Error finding container c64ff0759b4e0582da669df0d162803598a061dd7c8baa5aef06f79aba6470ac: Status 404 returned error can't find the container with id c64ff0759b4e0582da669df0d162803598a061dd7c8baa5aef06f79aba6470ac Dec 03 00:09:34 crc kubenswrapper[4953]: I1203 00:09:34.797895 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xfd2l" event={"ID":"52fd5773-3888-49f1-813d-10d262f4c901","Type":"ContainerStarted","Data":"a9095f2983a7ed1cb660c3459aaa9888337df9b6cd3750a5a83a8ccf53a7a70c"} Dec 03 00:09:34 crc kubenswrapper[4953]: I1203 00:09:34.803908 4953 generic.go:334] "Generic (PLEG): container finished" podID="3e5a99ac-9abf-4291-b773-fda0c4364959" containerID="9fbad7d01aef12b68dbe8856ab62492197d7d0237185915678fa26b4609f5e54" exitCode=0 Dec 03 00:09:34 crc kubenswrapper[4953]: I1203 00:09:34.804008 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ct49l" event={"ID":"3e5a99ac-9abf-4291-b773-fda0c4364959","Type":"ContainerDied","Data":"9fbad7d01aef12b68dbe8856ab62492197d7d0237185915678fa26b4609f5e54"} Dec 03 00:09:34 crc kubenswrapper[4953]: I1203 00:09:34.813426 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dq6nn" event={"ID":"9064a5ab-24ee-4258-8363-6ce4ffa54b3f","Type":"ContainerStarted","Data":"68ef5fb61b5fb8ed919f648e04de0009daebd9aed25ac57d75fe8c223f624424"} Dec 03 00:09:35 crc kubenswrapper[4953]: I1203 00:09:35.352855 4953 patch_prober.go:28] interesting pod/router-default-5444994796-r8l2l container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 03 00:09:35 crc kubenswrapper[4953]: [-]has-synced failed: reason withheld Dec 03 00:09:35 crc kubenswrapper[4953]: [+]process-running ok Dec 03 00:09:35 crc kubenswrapper[4953]: healthz check failed Dec 03 00:09:35 crc kubenswrapper[4953]: I1203 00:09:35.353332 4953 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-r8l2l" podUID="4edc57f1-5c0c-4108-bb3c-b2f104c6e08b" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 03 00:09:35 crc kubenswrapper[4953]: I1203 00:09:35.824079 4953 generic.go:334] "Generic (PLEG): container finished" podID="f17dcaea-cb22-4359-9480-e1d129cd309e" containerID="2514370e8db161da2ec1d92134a419afbf628145aa9c9a74b56340bc8b6ce59e" exitCode=0 Dec 03 00:09:35 crc kubenswrapper[4953]: I1203 00:09:35.824134 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-js4dj" event={"ID":"f17dcaea-cb22-4359-9480-e1d129cd309e","Type":"ContainerDied","Data":"2514370e8db161da2ec1d92134a419afbf628145aa9c9a74b56340bc8b6ce59e"} Dec 03 00:09:35 crc kubenswrapper[4953]: I1203 00:09:35.826234 4953 generic.go:334] "Generic (PLEG): container finished" podID="be2fa7e6-200e-44f0-9b40-57ed2d30db95" containerID="7933636e09868e01e20fa43288d4411f1cfd94ad47ec9d021d115adc318c4885" exitCode=0 Dec 03 00:09:35 crc kubenswrapper[4953]: I1203 00:09:35.826314 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-q94c6" event={"ID":"be2fa7e6-200e-44f0-9b40-57ed2d30db95","Type":"ContainerDied","Data":"7933636e09868e01e20fa43288d4411f1cfd94ad47ec9d021d115adc318c4885"} Dec 03 00:09:35 crc kubenswrapper[4953]: I1203 00:09:35.826342 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-q94c6" event={"ID":"be2fa7e6-200e-44f0-9b40-57ed2d30db95","Type":"ContainerStarted","Data":"c64ff0759b4e0582da669df0d162803598a061dd7c8baa5aef06f79aba6470ac"} Dec 03 00:09:35 crc kubenswrapper[4953]: I1203 00:09:35.826677 4953 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 03 00:09:35 crc kubenswrapper[4953]: I1203 00:09:35.829173 4953 generic.go:334] "Generic (PLEG): container finished" podID="9064a5ab-24ee-4258-8363-6ce4ffa54b3f" containerID="beeb25d3d42d4d66ef7fae9fc6aa9b0ae72ab4540161cf231a11bd225d262597" exitCode=0 Dec 03 00:09:35 crc kubenswrapper[4953]: I1203 00:09:35.829253 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dq6nn" event={"ID":"9064a5ab-24ee-4258-8363-6ce4ffa54b3f","Type":"ContainerDied","Data":"beeb25d3d42d4d66ef7fae9fc6aa9b0ae72ab4540161cf231a11bd225d262597"} Dec 03 00:09:35 crc kubenswrapper[4953]: I1203 00:09:35.833198 4953 generic.go:334] "Generic (PLEG): container finished" podID="6b353eef-b7f9-4dbc-b2aa-9860220438cc" containerID="261b04a7d566d558fe6b4892804ef231ee5a882005acbe560446cd4993e5fc9c" exitCode=0 Dec 03 00:09:35 crc kubenswrapper[4953]: I1203 00:09:35.833474 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hs6pg" event={"ID":"6b353eef-b7f9-4dbc-b2aa-9860220438cc","Type":"ContainerDied","Data":"261b04a7d566d558fe6b4892804ef231ee5a882005acbe560446cd4993e5fc9c"} Dec 03 00:09:35 crc kubenswrapper[4953]: I1203 00:09:35.845345 4953 generic.go:334] "Generic (PLEG): container finished" podID="e98b8740-f6fc-4db0-a807-664cf085391c" containerID="7033f14e4fa2cbc2578476d618efb3e44d904a9138d5b8ac8ec8d5d143af0f11" exitCode=0 Dec 03 00:09:35 crc kubenswrapper[4953]: I1203 00:09:35.845455 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5zxsj" event={"ID":"e98b8740-f6fc-4db0-a807-664cf085391c","Type":"ContainerDied","Data":"7033f14e4fa2cbc2578476d618efb3e44d904a9138d5b8ac8ec8d5d143af0f11"} Dec 03 00:09:35 crc kubenswrapper[4953]: I1203 00:09:35.848509 4953 generic.go:334] "Generic (PLEG): container finished" podID="52fd5773-3888-49f1-813d-10d262f4c901" containerID="a9095f2983a7ed1cb660c3459aaa9888337df9b6cd3750a5a83a8ccf53a7a70c" exitCode=0 Dec 03 00:09:35 crc kubenswrapper[4953]: I1203 00:09:35.848597 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xfd2l" event={"ID":"52fd5773-3888-49f1-813d-10d262f4c901","Type":"ContainerDied","Data":"a9095f2983a7ed1cb660c3459aaa9888337df9b6cd3750a5a83a8ccf53a7a70c"} Dec 03 00:09:35 crc kubenswrapper[4953]: I1203 00:09:35.859281 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"30fac6bb-4dfc-44c7-b8af-0342a8d504b7","Type":"ContainerStarted","Data":"3cff5e31fa38d0316936d05b5a29d2d5653944c856078abc1cd5cc5c1e541d7a"} Dec 03 00:09:35 crc kubenswrapper[4953]: I1203 00:09:35.866650 4953 generic.go:334] "Generic (PLEG): container finished" podID="496add53-6511-4560-82c5-3efeb65935f0" containerID="57759415d3b5338ee044069760976184c643ecf174ff7fdd85eea90738b2fdea" exitCode=0 Dec 03 00:09:35 crc kubenswrapper[4953]: I1203 00:09:35.866805 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6kz7m" event={"ID":"496add53-6511-4560-82c5-3efeb65935f0","Type":"ContainerDied","Data":"57759415d3b5338ee044069760976184c643ecf174ff7fdd85eea90738b2fdea"} Dec 03 00:09:35 crc kubenswrapper[4953]: I1203 00:09:35.868754 4953 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-697d97f7c8-t7ljb" Dec 03 00:09:36 crc kubenswrapper[4953]: I1203 00:09:36.075360 4953 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-697d97f7c8-t7ljb" podStartSLOduration=140.075332906 podStartE2EDuration="2m20.075332906s" podCreationTimestamp="2025-12-03 00:07:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 00:09:36.063418924 +0000 UTC m=+165.047699717" watchObservedRunningTime="2025-12-03 00:09:36.075332906 +0000 UTC m=+165.059613689" Dec 03 00:09:36 crc kubenswrapper[4953]: I1203 00:09:36.091298 4953 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="hostpath-provisioner/csi-hostpathplugin-2xt2r" podStartSLOduration=20.091278957 podStartE2EDuration="20.091278957s" podCreationTimestamp="2025-12-03 00:09:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 00:09:36.088289454 +0000 UTC m=+165.072570257" watchObservedRunningTime="2025-12-03 00:09:36.091278957 +0000 UTC m=+165.075559740" Dec 03 00:09:36 crc kubenswrapper[4953]: I1203 00:09:36.123383 4953 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/revision-pruner-9-crc" podStartSLOduration=5.123362493 podStartE2EDuration="5.123362493s" podCreationTimestamp="2025-12-03 00:09:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 00:09:36.104842149 +0000 UTC m=+165.089122932" watchObservedRunningTime="2025-12-03 00:09:36.123362493 +0000 UTC m=+165.107643276" Dec 03 00:09:36 crc kubenswrapper[4953]: I1203 00:09:36.207511 4953 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Dec 03 00:09:36 crc kubenswrapper[4953]: I1203 00:09:36.208161 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 03 00:09:36 crc kubenswrapper[4953]: I1203 00:09:36.211076 4953 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Dec 03 00:09:36 crc kubenswrapper[4953]: I1203 00:09:36.211077 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Dec 03 00:09:36 crc kubenswrapper[4953]: I1203 00:09:36.223916 4953 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Dec 03 00:09:36 crc kubenswrapper[4953]: I1203 00:09:36.309412 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/c75e311c-f334-4a7c-9f82-33c50c613f4b-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"c75e311c-f334-4a7c-9f82-33c50c613f4b\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 03 00:09:36 crc kubenswrapper[4953]: I1203 00:09:36.309543 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c75e311c-f334-4a7c-9f82-33c50c613f4b-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"c75e311c-f334-4a7c-9f82-33c50c613f4b\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 03 00:09:36 crc kubenswrapper[4953]: I1203 00:09:36.349145 4953 patch_prober.go:28] interesting pod/router-default-5444994796-r8l2l container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 03 00:09:36 crc kubenswrapper[4953]: [-]has-synced failed: reason withheld Dec 03 00:09:36 crc kubenswrapper[4953]: [+]process-running ok Dec 03 00:09:36 crc kubenswrapper[4953]: healthz check failed Dec 03 00:09:36 crc kubenswrapper[4953]: I1203 00:09:36.349193 4953 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-r8l2l" podUID="4edc57f1-5c0c-4108-bb3c-b2f104c6e08b" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 03 00:09:36 crc kubenswrapper[4953]: I1203 00:09:36.411197 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/c75e311c-f334-4a7c-9f82-33c50c613f4b-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"c75e311c-f334-4a7c-9f82-33c50c613f4b\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 03 00:09:36 crc kubenswrapper[4953]: I1203 00:09:36.411358 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c75e311c-f334-4a7c-9f82-33c50c613f4b-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"c75e311c-f334-4a7c-9f82-33c50c613f4b\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 03 00:09:36 crc kubenswrapper[4953]: I1203 00:09:36.411361 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/c75e311c-f334-4a7c-9f82-33c50c613f4b-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"c75e311c-f334-4a7c-9f82-33c50c613f4b\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 03 00:09:36 crc kubenswrapper[4953]: I1203 00:09:36.447587 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c75e311c-f334-4a7c-9f82-33c50c613f4b-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"c75e311c-f334-4a7c-9f82-33c50c613f4b\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 03 00:09:36 crc kubenswrapper[4953]: I1203 00:09:36.524353 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 03 00:09:37 crc kubenswrapper[4953]: I1203 00:09:37.080230 4953 generic.go:334] "Generic (PLEG): container finished" podID="30fac6bb-4dfc-44c7-b8af-0342a8d504b7" containerID="3cff5e31fa38d0316936d05b5a29d2d5653944c856078abc1cd5cc5c1e541d7a" exitCode=0 Dec 03 00:09:37 crc kubenswrapper[4953]: I1203 00:09:37.081626 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"30fac6bb-4dfc-44c7-b8af-0342a8d504b7","Type":"ContainerDied","Data":"3cff5e31fa38d0316936d05b5a29d2d5653944c856078abc1cd5cc5c1e541d7a"} Dec 03 00:09:37 crc kubenswrapper[4953]: I1203 00:09:37.338007 4953 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Dec 03 00:09:37 crc kubenswrapper[4953]: I1203 00:09:37.358285 4953 patch_prober.go:28] interesting pod/router-default-5444994796-r8l2l container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 03 00:09:37 crc kubenswrapper[4953]: [-]has-synced failed: reason withheld Dec 03 00:09:37 crc kubenswrapper[4953]: [+]process-running ok Dec 03 00:09:37 crc kubenswrapper[4953]: healthz check failed Dec 03 00:09:37 crc kubenswrapper[4953]: I1203 00:09:37.358343 4953 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-r8l2l" podUID="4edc57f1-5c0c-4108-bb3c-b2f104c6e08b" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 03 00:09:37 crc kubenswrapper[4953]: I1203 00:09:37.402819 4953 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-dns/dns-default-qbsjg" Dec 03 00:09:38 crc kubenswrapper[4953]: I1203 00:09:38.112664 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"c75e311c-f334-4a7c-9f82-33c50c613f4b","Type":"ContainerStarted","Data":"4a2d36a8f848e5920e54afd6971f42e2fe6118d177e465c23b46b6851aedd8d0"} Dec 03 00:09:38 crc kubenswrapper[4953]: I1203 00:09:38.197259 4953 patch_prober.go:28] interesting pod/downloads-7954f5f757-5rzxr container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" start-of-body= Dec 03 00:09:38 crc kubenswrapper[4953]: I1203 00:09:38.197305 4953 patch_prober.go:28] interesting pod/downloads-7954f5f757-5rzxr container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" start-of-body= Dec 03 00:09:38 crc kubenswrapper[4953]: I1203 00:09:38.197693 4953 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-5rzxr" podUID="e0e3abc1-91d2-4a6a-aa1c-dc166f30d75a" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" Dec 03 00:09:38 crc kubenswrapper[4953]: I1203 00:09:38.197712 4953 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-5rzxr" podUID="e0e3abc1-91d2-4a6a-aa1c-dc166f30d75a" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" Dec 03 00:09:38 crc kubenswrapper[4953]: I1203 00:09:38.422195 4953 patch_prober.go:28] interesting pod/router-default-5444994796-r8l2l container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 03 00:09:38 crc kubenswrapper[4953]: [-]has-synced failed: reason withheld Dec 03 00:09:38 crc kubenswrapper[4953]: [+]process-running ok Dec 03 00:09:38 crc kubenswrapper[4953]: healthz check failed Dec 03 00:09:38 crc kubenswrapper[4953]: I1203 00:09:38.422255 4953 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-r8l2l" podUID="4edc57f1-5c0c-4108-bb3c-b2f104c6e08b" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 03 00:09:39 crc kubenswrapper[4953]: I1203 00:09:39.118405 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"c75e311c-f334-4a7c-9f82-33c50c613f4b","Type":"ContainerStarted","Data":"826888ed1f764889744f3d2b923986ce3122b93a14329d085876d08ba6d45470"} Dec 03 00:09:39 crc kubenswrapper[4953]: I1203 00:09:39.154140 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/bd03a758-0ca9-47d8-86f7-18b2d610df43-metrics-certs\") pod \"network-metrics-daemon-xshk4\" (UID: \"bd03a758-0ca9-47d8-86f7-18b2d610df43\") " pod="openshift-multus/network-metrics-daemon-xshk4" Dec 03 00:09:39 crc kubenswrapper[4953]: I1203 00:09:39.174787 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/bd03a758-0ca9-47d8-86f7-18b2d610df43-metrics-certs\") pod \"network-metrics-daemon-xshk4\" (UID: \"bd03a758-0ca9-47d8-86f7-18b2d610df43\") " pod="openshift-multus/network-metrics-daemon-xshk4" Dec 03 00:09:39 crc kubenswrapper[4953]: I1203 00:09:39.187954 4953 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/revision-pruner-8-crc" podStartSLOduration=3.18793326 podStartE2EDuration="3.18793326s" podCreationTimestamp="2025-12-03 00:09:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 00:09:39.184153028 +0000 UTC m=+168.168433811" watchObservedRunningTime="2025-12-03 00:09:39.18793326 +0000 UTC m=+168.172214043" Dec 03 00:09:39 crc kubenswrapper[4953]: I1203 00:09:39.205556 4953 patch_prober.go:28] interesting pod/console-f9d7485db-m2msx container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.16:8443/health\": dial tcp 10.217.0.16:8443: connect: connection refused" start-of-body= Dec 03 00:09:39 crc kubenswrapper[4953]: I1203 00:09:39.205610 4953 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-m2msx" podUID="4ce6dbe8-0292-4172-81a3-f5cdfdba6120" containerName="console" probeResult="failure" output="Get \"https://10.217.0.16:8443/health\": dial tcp 10.217.0.16:8443: connect: connection refused" Dec 03 00:09:39 crc kubenswrapper[4953]: I1203 00:09:39.401525 4953 patch_prober.go:28] interesting pod/router-default-5444994796-r8l2l container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 03 00:09:39 crc kubenswrapper[4953]: [-]has-synced failed: reason withheld Dec 03 00:09:39 crc kubenswrapper[4953]: [+]process-running ok Dec 03 00:09:39 crc kubenswrapper[4953]: healthz check failed Dec 03 00:09:39 crc kubenswrapper[4953]: I1203 00:09:39.401580 4953 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-r8l2l" podUID="4edc57f1-5c0c-4108-bb3c-b2f104c6e08b" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 03 00:09:39 crc kubenswrapper[4953]: I1203 00:09:39.468943 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xshk4" Dec 03 00:09:39 crc kubenswrapper[4953]: I1203 00:09:39.481215 4953 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 03 00:09:39 crc kubenswrapper[4953]: I1203 00:09:39.501096 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/30fac6bb-4dfc-44c7-b8af-0342a8d504b7-kubelet-dir\") pod \"30fac6bb-4dfc-44c7-b8af-0342a8d504b7\" (UID: \"30fac6bb-4dfc-44c7-b8af-0342a8d504b7\") " Dec 03 00:09:39 crc kubenswrapper[4953]: I1203 00:09:39.501254 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/30fac6bb-4dfc-44c7-b8af-0342a8d504b7-kube-api-access\") pod \"30fac6bb-4dfc-44c7-b8af-0342a8d504b7\" (UID: \"30fac6bb-4dfc-44c7-b8af-0342a8d504b7\") " Dec 03 00:09:39 crc kubenswrapper[4953]: I1203 00:09:39.503140 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/30fac6bb-4dfc-44c7-b8af-0342a8d504b7-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "30fac6bb-4dfc-44c7-b8af-0342a8d504b7" (UID: "30fac6bb-4dfc-44c7-b8af-0342a8d504b7"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 00:09:39 crc kubenswrapper[4953]: I1203 00:09:39.537545 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/30fac6bb-4dfc-44c7-b8af-0342a8d504b7-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "30fac6bb-4dfc-44c7-b8af-0342a8d504b7" (UID: "30fac6bb-4dfc-44c7-b8af-0342a8d504b7"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:09:39 crc kubenswrapper[4953]: I1203 00:09:39.603556 4953 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/30fac6bb-4dfc-44c7-b8af-0342a8d504b7-kubelet-dir\") on node \"crc\" DevicePath \"\"" Dec 03 00:09:39 crc kubenswrapper[4953]: I1203 00:09:39.603577 4953 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/30fac6bb-4dfc-44c7-b8af-0342a8d504b7-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 03 00:09:39 crc kubenswrapper[4953]: I1203 00:09:39.645789 4953 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-kx9dt" Dec 03 00:09:40 crc kubenswrapper[4953]: I1203 00:09:40.131150 4953 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 03 00:09:40 crc kubenswrapper[4953]: I1203 00:09:40.131753 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"30fac6bb-4dfc-44c7-b8af-0342a8d504b7","Type":"ContainerDied","Data":"c6f1decaff066a9700bed15273ab687fcdc3ae7bb81fd359a8ba51e570df631a"} Dec 03 00:09:40 crc kubenswrapper[4953]: I1203 00:09:40.131780 4953 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c6f1decaff066a9700bed15273ab687fcdc3ae7bb81fd359a8ba51e570df631a" Dec 03 00:09:40 crc kubenswrapper[4953]: I1203 00:09:40.348564 4953 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-xshk4"] Dec 03 00:09:40 crc kubenswrapper[4953]: I1203 00:09:40.351877 4953 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-ingress/router-default-5444994796-r8l2l" Dec 03 00:09:40 crc kubenswrapper[4953]: I1203 00:09:40.365654 4953 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ingress/router-default-5444994796-r8l2l" Dec 03 00:09:41 crc kubenswrapper[4953]: I1203 00:09:41.183878 4953 generic.go:334] "Generic (PLEG): container finished" podID="c75e311c-f334-4a7c-9f82-33c50c613f4b" containerID="826888ed1f764889744f3d2b923986ce3122b93a14329d085876d08ba6d45470" exitCode=0 Dec 03 00:09:41 crc kubenswrapper[4953]: I1203 00:09:41.184041 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"c75e311c-f334-4a7c-9f82-33c50c613f4b","Type":"ContainerDied","Data":"826888ed1f764889744f3d2b923986ce3122b93a14329d085876d08ba6d45470"} Dec 03 00:09:41 crc kubenswrapper[4953]: I1203 00:09:41.205443 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-xshk4" event={"ID":"bd03a758-0ca9-47d8-86f7-18b2d610df43","Type":"ContainerStarted","Data":"d503f0f3c29bc6bdaee6e57d25b217295a11c63eb4ce84121081b530f86a2a26"} Dec 03 00:09:41 crc kubenswrapper[4953]: I1203 00:09:41.205544 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-xshk4" event={"ID":"bd03a758-0ca9-47d8-86f7-18b2d610df43","Type":"ContainerStarted","Data":"05713170f806f4223b6401a6ef5decb6419a77cea0fb2725afa546df1d12cecc"} Dec 03 00:09:43 crc kubenswrapper[4953]: I1203 00:09:43.301705 4953 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 03 00:09:43 crc kubenswrapper[4953]: I1203 00:09:43.337841 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"c75e311c-f334-4a7c-9f82-33c50c613f4b","Type":"ContainerDied","Data":"4a2d36a8f848e5920e54afd6971f42e2fe6118d177e465c23b46b6851aedd8d0"} Dec 03 00:09:43 crc kubenswrapper[4953]: I1203 00:09:43.338010 4953 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4a2d36a8f848e5920e54afd6971f42e2fe6118d177e465c23b46b6851aedd8d0" Dec 03 00:09:43 crc kubenswrapper[4953]: I1203 00:09:43.338102 4953 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 03 00:09:43 crc kubenswrapper[4953]: I1203 00:09:43.351668 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-xshk4" event={"ID":"bd03a758-0ca9-47d8-86f7-18b2d610df43","Type":"ContainerStarted","Data":"85e4c27ab85ef389e38db3ae069c8fd5bd0c615f03ecf54ba4b4a36622867d10"} Dec 03 00:09:43 crc kubenswrapper[4953]: I1203 00:09:43.518509 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/c75e311c-f334-4a7c-9f82-33c50c613f4b-kubelet-dir\") pod \"c75e311c-f334-4a7c-9f82-33c50c613f4b\" (UID: \"c75e311c-f334-4a7c-9f82-33c50c613f4b\") " Dec 03 00:09:43 crc kubenswrapper[4953]: I1203 00:09:43.519003 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c75e311c-f334-4a7c-9f82-33c50c613f4b-kube-api-access\") pod \"c75e311c-f334-4a7c-9f82-33c50c613f4b\" (UID: \"c75e311c-f334-4a7c-9f82-33c50c613f4b\") " Dec 03 00:09:43 crc kubenswrapper[4953]: I1203 00:09:43.518735 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c75e311c-f334-4a7c-9f82-33c50c613f4b-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "c75e311c-f334-4a7c-9f82-33c50c613f4b" (UID: "c75e311c-f334-4a7c-9f82-33c50c613f4b"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 00:09:43 crc kubenswrapper[4953]: I1203 00:09:43.676967 4953 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/c75e311c-f334-4a7c-9f82-33c50c613f4b-kubelet-dir\") on node \"crc\" DevicePath \"\"" Dec 03 00:09:43 crc kubenswrapper[4953]: I1203 00:09:43.729537 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c75e311c-f334-4a7c-9f82-33c50c613f4b-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "c75e311c-f334-4a7c-9f82-33c50c613f4b" (UID: "c75e311c-f334-4a7c-9f82-33c50c613f4b"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:09:43 crc kubenswrapper[4953]: I1203 00:09:43.781111 4953 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c75e311c-f334-4a7c-9f82-33c50c613f4b-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 03 00:09:48 crc kubenswrapper[4953]: I1203 00:09:48.214008 4953 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/downloads-7954f5f757-5rzxr" Dec 03 00:09:48 crc kubenswrapper[4953]: I1203 00:09:48.264294 4953 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/network-metrics-daemon-xshk4" podStartSLOduration=152.264255231 podStartE2EDuration="2m32.264255231s" podCreationTimestamp="2025-12-03 00:07:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 00:09:43.46632353 +0000 UTC m=+172.450604323" watchObservedRunningTime="2025-12-03 00:09:48.264255231 +0000 UTC m=+177.248536014" Dec 03 00:09:48 crc kubenswrapper[4953]: I1203 00:09:48.944944 4953 patch_prober.go:28] interesting pod/machine-config-daemon-6dfbn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 00:09:48 crc kubenswrapper[4953]: I1203 00:09:48.945393 4953 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-6dfbn" podUID="53ffb1e9-9a2e-4869-80d7-776f4aa7ef8b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 00:09:49 crc kubenswrapper[4953]: I1203 00:09:49.237354 4953 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-f9d7485db-m2msx" Dec 03 00:09:49 crc kubenswrapper[4953]: I1203 00:09:49.246095 4953 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-f9d7485db-m2msx" Dec 03 00:09:51 crc kubenswrapper[4953]: I1203 00:09:51.498389 4953 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-697d97f7c8-t7ljb" Dec 03 00:09:55 crc kubenswrapper[4953]: I1203 00:09:55.788249 4953 generic.go:334] "Generic (PLEG): container finished" podID="a68720c3-4f74-4435-91c6-99fd1b561416" containerID="ce83851e48b19dc51f2a638a8c17cff4505f4f43bd73cda58510bf0e9d803ffa" exitCode=0 Dec 03 00:09:55 crc kubenswrapper[4953]: I1203 00:09:55.788327 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-pruner-29412000-njkvg" event={"ID":"a68720c3-4f74-4435-91c6-99fd1b561416","Type":"ContainerDied","Data":"ce83851e48b19dc51f2a638a8c17cff4505f4f43bd73cda58510bf0e9d803ffa"} Dec 03 00:09:58 crc kubenswrapper[4953]: I1203 00:09:58.854840 4953 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 00:09:58 crc kubenswrapper[4953]: I1203 00:09:58.981359 4953 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-dlqdp" Dec 03 00:10:09 crc kubenswrapper[4953]: I1203 00:10:09.417330 4953 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Dec 03 00:10:09 crc kubenswrapper[4953]: E1203 00:10:09.418447 4953 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c75e311c-f334-4a7c-9f82-33c50c613f4b" containerName="pruner" Dec 03 00:10:09 crc kubenswrapper[4953]: I1203 00:10:09.418469 4953 state_mem.go:107] "Deleted CPUSet assignment" podUID="c75e311c-f334-4a7c-9f82-33c50c613f4b" containerName="pruner" Dec 03 00:10:09 crc kubenswrapper[4953]: E1203 00:10:09.418496 4953 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="30fac6bb-4dfc-44c7-b8af-0342a8d504b7" containerName="pruner" Dec 03 00:10:09 crc kubenswrapper[4953]: I1203 00:10:09.418532 4953 state_mem.go:107] "Deleted CPUSet assignment" podUID="30fac6bb-4dfc-44c7-b8af-0342a8d504b7" containerName="pruner" Dec 03 00:10:09 crc kubenswrapper[4953]: I1203 00:10:09.418712 4953 memory_manager.go:354] "RemoveStaleState removing state" podUID="c75e311c-f334-4a7c-9f82-33c50c613f4b" containerName="pruner" Dec 03 00:10:09 crc kubenswrapper[4953]: I1203 00:10:09.418735 4953 memory_manager.go:354] "RemoveStaleState removing state" podUID="30fac6bb-4dfc-44c7-b8af-0342a8d504b7" containerName="pruner" Dec 03 00:10:09 crc kubenswrapper[4953]: I1203 00:10:09.419487 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 03 00:10:09 crc kubenswrapper[4953]: I1203 00:10:09.429598 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Dec 03 00:10:09 crc kubenswrapper[4953]: I1203 00:10:09.430175 4953 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Dec 03 00:10:09 crc kubenswrapper[4953]: I1203 00:10:09.434653 4953 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Dec 03 00:10:09 crc kubenswrapper[4953]: I1203 00:10:09.506927 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/622700d5-0094-4bd6-b12a-8fb43581b5ab-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"622700d5-0094-4bd6-b12a-8fb43581b5ab\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 03 00:10:09 crc kubenswrapper[4953]: I1203 00:10:09.507010 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/622700d5-0094-4bd6-b12a-8fb43581b5ab-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"622700d5-0094-4bd6-b12a-8fb43581b5ab\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 03 00:10:09 crc kubenswrapper[4953]: I1203 00:10:09.608313 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/622700d5-0094-4bd6-b12a-8fb43581b5ab-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"622700d5-0094-4bd6-b12a-8fb43581b5ab\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 03 00:10:09 crc kubenswrapper[4953]: I1203 00:10:09.608407 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/622700d5-0094-4bd6-b12a-8fb43581b5ab-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"622700d5-0094-4bd6-b12a-8fb43581b5ab\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 03 00:10:09 crc kubenswrapper[4953]: I1203 00:10:09.608628 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/622700d5-0094-4bd6-b12a-8fb43581b5ab-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"622700d5-0094-4bd6-b12a-8fb43581b5ab\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 03 00:10:09 crc kubenswrapper[4953]: I1203 00:10:09.616965 4953 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-pruner-29412000-njkvg" Dec 03 00:10:09 crc kubenswrapper[4953]: I1203 00:10:09.637352 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/622700d5-0094-4bd6-b12a-8fb43581b5ab-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"622700d5-0094-4bd6-b12a-8fb43581b5ab\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 03 00:10:09 crc kubenswrapper[4953]: I1203 00:10:09.709040 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/a68720c3-4f74-4435-91c6-99fd1b561416-serviceca\") pod \"a68720c3-4f74-4435-91c6-99fd1b561416\" (UID: \"a68720c3-4f74-4435-91c6-99fd1b561416\") " Dec 03 00:10:09 crc kubenswrapper[4953]: I1203 00:10:09.709117 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-67dx2\" (UniqueName: \"kubernetes.io/projected/a68720c3-4f74-4435-91c6-99fd1b561416-kube-api-access-67dx2\") pod \"a68720c3-4f74-4435-91c6-99fd1b561416\" (UID: \"a68720c3-4f74-4435-91c6-99fd1b561416\") " Dec 03 00:10:09 crc kubenswrapper[4953]: I1203 00:10:09.711696 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a68720c3-4f74-4435-91c6-99fd1b561416-serviceca" (OuterVolumeSpecName: "serviceca") pod "a68720c3-4f74-4435-91c6-99fd1b561416" (UID: "a68720c3-4f74-4435-91c6-99fd1b561416"). InnerVolumeSpecName "serviceca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:10:09 crc kubenswrapper[4953]: I1203 00:10:09.713552 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a68720c3-4f74-4435-91c6-99fd1b561416-kube-api-access-67dx2" (OuterVolumeSpecName: "kube-api-access-67dx2") pod "a68720c3-4f74-4435-91c6-99fd1b561416" (UID: "a68720c3-4f74-4435-91c6-99fd1b561416"). InnerVolumeSpecName "kube-api-access-67dx2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:10:09 crc kubenswrapper[4953]: I1203 00:10:09.748051 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 03 00:10:09 crc kubenswrapper[4953]: I1203 00:10:09.810256 4953 reconciler_common.go:293] "Volume detached for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/a68720c3-4f74-4435-91c6-99fd1b561416-serviceca\") on node \"crc\" DevicePath \"\"" Dec 03 00:10:09 crc kubenswrapper[4953]: I1203 00:10:09.810303 4953 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-67dx2\" (UniqueName: \"kubernetes.io/projected/a68720c3-4f74-4435-91c6-99fd1b561416-kube-api-access-67dx2\") on node \"crc\" DevicePath \"\"" Dec 03 00:10:09 crc kubenswrapper[4953]: I1203 00:10:09.998474 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-pruner-29412000-njkvg" event={"ID":"a68720c3-4f74-4435-91c6-99fd1b561416","Type":"ContainerDied","Data":"a046b8835dc4b76faedddf1622c0469842261ac7a0761bfc7b1b596ad6b7814c"} Dec 03 00:10:09 crc kubenswrapper[4953]: I1203 00:10:09.998600 4953 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a046b8835dc4b76faedddf1622c0469842261ac7a0761bfc7b1b596ad6b7814c" Dec 03 00:10:09 crc kubenswrapper[4953]: I1203 00:10:09.998614 4953 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-pruner-29412000-njkvg" Dec 03 00:10:14 crc kubenswrapper[4953]: I1203 00:10:14.003823 4953 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Dec 03 00:10:14 crc kubenswrapper[4953]: E1203 00:10:14.004423 4953 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a68720c3-4f74-4435-91c6-99fd1b561416" containerName="image-pruner" Dec 03 00:10:14 crc kubenswrapper[4953]: I1203 00:10:14.004438 4953 state_mem.go:107] "Deleted CPUSet assignment" podUID="a68720c3-4f74-4435-91c6-99fd1b561416" containerName="image-pruner" Dec 03 00:10:14 crc kubenswrapper[4953]: I1203 00:10:14.004560 4953 memory_manager.go:354] "RemoveStaleState removing state" podUID="a68720c3-4f74-4435-91c6-99fd1b561416" containerName="image-pruner" Dec 03 00:10:14 crc kubenswrapper[4953]: I1203 00:10:14.004903 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Dec 03 00:10:14 crc kubenswrapper[4953]: I1203 00:10:14.160952 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/bc833625-0910-40d3-8f2d-801b4b9c40d2-var-lock\") pod \"installer-9-crc\" (UID: \"bc833625-0910-40d3-8f2d-801b4b9c40d2\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 03 00:10:14 crc kubenswrapper[4953]: I1203 00:10:14.161034 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/bc833625-0910-40d3-8f2d-801b4b9c40d2-kubelet-dir\") pod \"installer-9-crc\" (UID: \"bc833625-0910-40d3-8f2d-801b4b9c40d2\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 03 00:10:14 crc kubenswrapper[4953]: I1203 00:10:14.161106 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/bc833625-0910-40d3-8f2d-801b4b9c40d2-kube-api-access\") pod \"installer-9-crc\" (UID: \"bc833625-0910-40d3-8f2d-801b4b9c40d2\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 03 00:10:14 crc kubenswrapper[4953]: I1203 00:10:14.172285 4953 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Dec 03 00:10:14 crc kubenswrapper[4953]: I1203 00:10:14.262078 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/bc833625-0910-40d3-8f2d-801b4b9c40d2-var-lock\") pod \"installer-9-crc\" (UID: \"bc833625-0910-40d3-8f2d-801b4b9c40d2\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 03 00:10:14 crc kubenswrapper[4953]: I1203 00:10:14.262223 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/bc833625-0910-40d3-8f2d-801b4b9c40d2-kubelet-dir\") pod \"installer-9-crc\" (UID: \"bc833625-0910-40d3-8f2d-801b4b9c40d2\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 03 00:10:14 crc kubenswrapper[4953]: I1203 00:10:14.262171 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/bc833625-0910-40d3-8f2d-801b4b9c40d2-var-lock\") pod \"installer-9-crc\" (UID: \"bc833625-0910-40d3-8f2d-801b4b9c40d2\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 03 00:10:14 crc kubenswrapper[4953]: I1203 00:10:14.262341 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/bc833625-0910-40d3-8f2d-801b4b9c40d2-kube-api-access\") pod \"installer-9-crc\" (UID: \"bc833625-0910-40d3-8f2d-801b4b9c40d2\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 03 00:10:14 crc kubenswrapper[4953]: I1203 00:10:14.262414 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/bc833625-0910-40d3-8f2d-801b4b9c40d2-kubelet-dir\") pod \"installer-9-crc\" (UID: \"bc833625-0910-40d3-8f2d-801b4b9c40d2\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 03 00:10:14 crc kubenswrapper[4953]: I1203 00:10:14.294884 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/bc833625-0910-40d3-8f2d-801b4b9c40d2-kube-api-access\") pod \"installer-9-crc\" (UID: \"bc833625-0910-40d3-8f2d-801b4b9c40d2\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 03 00:10:14 crc kubenswrapper[4953]: I1203 00:10:14.475672 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Dec 03 00:10:18 crc kubenswrapper[4953]: I1203 00:10:18.944826 4953 patch_prober.go:28] interesting pod/machine-config-daemon-6dfbn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 00:10:18 crc kubenswrapper[4953]: I1203 00:10:18.945845 4953 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-6dfbn" podUID="53ffb1e9-9a2e-4869-80d7-776f4aa7ef8b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 00:10:18 crc kubenswrapper[4953]: I1203 00:10:18.945943 4953 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-6dfbn" Dec 03 00:10:18 crc kubenswrapper[4953]: I1203 00:10:18.949055 4953 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"7df84405eda520ad5a22d35750edf07ef3c3977a756ad5739c54f91cba9da7b2"} pod="openshift-machine-config-operator/machine-config-daemon-6dfbn" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 03 00:10:18 crc kubenswrapper[4953]: I1203 00:10:18.949236 4953 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-6dfbn" podUID="53ffb1e9-9a2e-4869-80d7-776f4aa7ef8b" containerName="machine-config-daemon" containerID="cri-o://7df84405eda520ad5a22d35750edf07ef3c3977a756ad5739c54f91cba9da7b2" gracePeriod=600 Dec 03 00:10:22 crc kubenswrapper[4953]: I1203 00:10:22.246286 4953 generic.go:334] "Generic (PLEG): container finished" podID="53ffb1e9-9a2e-4869-80d7-776f4aa7ef8b" containerID="7df84405eda520ad5a22d35750edf07ef3c3977a756ad5739c54f91cba9da7b2" exitCode=0 Dec 03 00:10:22 crc kubenswrapper[4953]: I1203 00:10:22.246339 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-6dfbn" event={"ID":"53ffb1e9-9a2e-4869-80d7-776f4aa7ef8b","Type":"ContainerDied","Data":"7df84405eda520ad5a22d35750edf07ef3c3977a756ad5739c54f91cba9da7b2"} Dec 03 00:10:31 crc kubenswrapper[4953]: E1203 00:10:31.620922 4953 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Dec 03 00:10:31 crc kubenswrapper[4953]: E1203 00:10:31.621691 4953 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-jfmxh,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-js4dj_openshift-marketplace(f17dcaea-cb22-4359-9480-e1d129cd309e): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 03 00:10:31 crc kubenswrapper[4953]: E1203 00:10:31.622999 4953 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-js4dj" podUID="f17dcaea-cb22-4359-9480-e1d129cd309e" Dec 03 00:10:40 crc kubenswrapper[4953]: E1203 00:10:40.647175 4953 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-js4dj" podUID="f17dcaea-cb22-4359-9480-e1d129cd309e" Dec 03 00:10:41 crc kubenswrapper[4953]: E1203 00:10:41.354606 4953 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Dec 03 00:10:41 crc kubenswrapper[4953]: E1203 00:10:41.355334 4953 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-ghrbv,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-q94c6_openshift-marketplace(be2fa7e6-200e-44f0-9b40-57ed2d30db95): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 03 00:10:41 crc kubenswrapper[4953]: E1203 00:10:41.356509 4953 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-q94c6" podUID="be2fa7e6-200e-44f0-9b40-57ed2d30db95" Dec 03 00:10:43 crc kubenswrapper[4953]: E1203 00:10:43.889992 4953 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-q94c6" podUID="be2fa7e6-200e-44f0-9b40-57ed2d30db95" Dec 03 00:10:44 crc kubenswrapper[4953]: E1203 00:10:44.746533 4953 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Dec 03 00:10:44 crc kubenswrapper[4953]: E1203 00:10:44.746941 4953 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-kv2jf,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-5zxsj_openshift-marketplace(e98b8740-f6fc-4db0-a807-664cf085391c): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 03 00:10:44 crc kubenswrapper[4953]: E1203 00:10:44.748419 4953 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-5zxsj" podUID="e98b8740-f6fc-4db0-a807-664cf085391c" Dec 03 00:10:47 crc kubenswrapper[4953]: E1203 00:10:47.198542 4953 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-5zxsj" podUID="e98b8740-f6fc-4db0-a807-664cf085391c" Dec 03 00:10:47 crc kubenswrapper[4953]: E1203 00:10:47.315777 4953 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Dec 03 00:10:47 crc kubenswrapper[4953]: E1203 00:10:47.316439 4953 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-m89m9,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-hs6pg_openshift-marketplace(6b353eef-b7f9-4dbc-b2aa-9860220438cc): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 03 00:10:47 crc kubenswrapper[4953]: E1203 00:10:47.318421 4953 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-hs6pg" podUID="6b353eef-b7f9-4dbc-b2aa-9860220438cc" Dec 03 00:10:47 crc kubenswrapper[4953]: E1203 00:10:47.460925 4953 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Dec 03 00:10:47 crc kubenswrapper[4953]: E1203 00:10:47.461419 4953 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-lffhw,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-ct49l_openshift-marketplace(3e5a99ac-9abf-4291-b773-fda0c4364959): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 03 00:10:47 crc kubenswrapper[4953]: E1203 00:10:47.461698 4953 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Dec 03 00:10:47 crc kubenswrapper[4953]: E1203 00:10:47.461755 4953 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-gjvz8,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-6kz7m_openshift-marketplace(496add53-6511-4560-82c5-3efeb65935f0): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 03 00:10:47 crc kubenswrapper[4953]: E1203 00:10:47.463234 4953 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-ct49l" podUID="3e5a99ac-9abf-4291-b773-fda0c4364959" Dec 03 00:10:47 crc kubenswrapper[4953]: E1203 00:10:47.463254 4953 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-6kz7m" podUID="496add53-6511-4560-82c5-3efeb65935f0" Dec 03 00:10:47 crc kubenswrapper[4953]: E1203 00:10:47.472629 4953 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-hs6pg" podUID="6b353eef-b7f9-4dbc-b2aa-9860220438cc" Dec 03 00:10:47 crc kubenswrapper[4953]: I1203 00:10:47.532553 4953 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Dec 03 00:10:47 crc kubenswrapper[4953]: I1203 00:10:47.695141 4953 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Dec 03 00:10:47 crc kubenswrapper[4953]: E1203 00:10:47.789434 4953 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Dec 03 00:10:47 crc kubenswrapper[4953]: E1203 00:10:47.790366 4953 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-qlmtg,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-xfd2l_openshift-marketplace(52fd5773-3888-49f1-813d-10d262f4c901): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 03 00:10:47 crc kubenswrapper[4953]: E1203 00:10:47.792036 4953 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-xfd2l" podUID="52fd5773-3888-49f1-813d-10d262f4c901" Dec 03 00:10:47 crc kubenswrapper[4953]: E1203 00:10:47.853102 4953 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Dec 03 00:10:47 crc kubenswrapper[4953]: E1203 00:10:47.853279 4953 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-8d5tx,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-dq6nn_openshift-marketplace(9064a5ab-24ee-4258-8363-6ce4ffa54b3f): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 03 00:10:47 crc kubenswrapper[4953]: E1203 00:10:47.855838 4953 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-dq6nn" podUID="9064a5ab-24ee-4258-8363-6ce4ffa54b3f" Dec 03 00:10:48 crc kubenswrapper[4953]: I1203 00:10:48.475086 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-6dfbn" event={"ID":"53ffb1e9-9a2e-4869-80d7-776f4aa7ef8b","Type":"ContainerStarted","Data":"cd3fdc5de406f82703cc48138f69373c50d1862f3fb1f4cc4572aa863d2eca4c"} Dec 03 00:10:48 crc kubenswrapper[4953]: I1203 00:10:48.483121 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"622700d5-0094-4bd6-b12a-8fb43581b5ab","Type":"ContainerStarted","Data":"e924592094d5726a10b54788f127cd0a9f2de98ec2c13632cc050116b7b39b34"} Dec 03 00:10:48 crc kubenswrapper[4953]: I1203 00:10:48.483188 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"622700d5-0094-4bd6-b12a-8fb43581b5ab","Type":"ContainerStarted","Data":"93b066fdbbc97d1257f39b14356258c5c35804fe30647b77af9b6c107ceacee7"} Dec 03 00:10:48 crc kubenswrapper[4953]: I1203 00:10:48.485663 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"bc833625-0910-40d3-8f2d-801b4b9c40d2","Type":"ContainerStarted","Data":"bebd17b4416c9e8f950db20eecd3806544c7aa8c0f46296470a6535c84598c5a"} Dec 03 00:10:48 crc kubenswrapper[4953]: I1203 00:10:48.485734 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"bc833625-0910-40d3-8f2d-801b4b9c40d2","Type":"ContainerStarted","Data":"e4ef29be021e5aec33a5212e8eb58d157536c4efd345782f17b4cbd4069a3cb7"} Dec 03 00:10:48 crc kubenswrapper[4953]: E1203 00:10:48.490529 4953 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-6kz7m" podUID="496add53-6511-4560-82c5-3efeb65935f0" Dec 03 00:10:48 crc kubenswrapper[4953]: E1203 00:10:48.490780 4953 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-dq6nn" podUID="9064a5ab-24ee-4258-8363-6ce4ffa54b3f" Dec 03 00:10:48 crc kubenswrapper[4953]: E1203 00:10:48.490851 4953 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-xfd2l" podUID="52fd5773-3888-49f1-813d-10d262f4c901" Dec 03 00:10:48 crc kubenswrapper[4953]: I1203 00:10:48.556870 4953 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/installer-9-crc" podStartSLOduration=35.556847684 podStartE2EDuration="35.556847684s" podCreationTimestamp="2025-12-03 00:10:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 00:10:48.535189671 +0000 UTC m=+237.519470454" watchObservedRunningTime="2025-12-03 00:10:48.556847684 +0000 UTC m=+237.541128467" Dec 03 00:10:50 crc kubenswrapper[4953]: I1203 00:10:50.496329 4953 generic.go:334] "Generic (PLEG): container finished" podID="622700d5-0094-4bd6-b12a-8fb43581b5ab" containerID="e924592094d5726a10b54788f127cd0a9f2de98ec2c13632cc050116b7b39b34" exitCode=0 Dec 03 00:10:50 crc kubenswrapper[4953]: I1203 00:10:50.496375 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"622700d5-0094-4bd6-b12a-8fb43581b5ab","Type":"ContainerDied","Data":"e924592094d5726a10b54788f127cd0a9f2de98ec2c13632cc050116b7b39b34"} Dec 03 00:10:51 crc kubenswrapper[4953]: I1203 00:10:51.815645 4953 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 03 00:10:51 crc kubenswrapper[4953]: I1203 00:10:51.920564 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/622700d5-0094-4bd6-b12a-8fb43581b5ab-kube-api-access\") pod \"622700d5-0094-4bd6-b12a-8fb43581b5ab\" (UID: \"622700d5-0094-4bd6-b12a-8fb43581b5ab\") " Dec 03 00:10:51 crc kubenswrapper[4953]: I1203 00:10:51.920647 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/622700d5-0094-4bd6-b12a-8fb43581b5ab-kubelet-dir\") pod \"622700d5-0094-4bd6-b12a-8fb43581b5ab\" (UID: \"622700d5-0094-4bd6-b12a-8fb43581b5ab\") " Dec 03 00:10:51 crc kubenswrapper[4953]: I1203 00:10:51.921026 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/622700d5-0094-4bd6-b12a-8fb43581b5ab-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "622700d5-0094-4bd6-b12a-8fb43581b5ab" (UID: "622700d5-0094-4bd6-b12a-8fb43581b5ab"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 00:10:51 crc kubenswrapper[4953]: I1203 00:10:51.927712 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/622700d5-0094-4bd6-b12a-8fb43581b5ab-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "622700d5-0094-4bd6-b12a-8fb43581b5ab" (UID: "622700d5-0094-4bd6-b12a-8fb43581b5ab"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:10:52 crc kubenswrapper[4953]: I1203 00:10:52.022193 4953 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/622700d5-0094-4bd6-b12a-8fb43581b5ab-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 03 00:10:52 crc kubenswrapper[4953]: I1203 00:10:52.022236 4953 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/622700d5-0094-4bd6-b12a-8fb43581b5ab-kubelet-dir\") on node \"crc\" DevicePath \"\"" Dec 03 00:10:52 crc kubenswrapper[4953]: I1203 00:10:52.507993 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"622700d5-0094-4bd6-b12a-8fb43581b5ab","Type":"ContainerDied","Data":"93b066fdbbc97d1257f39b14356258c5c35804fe30647b77af9b6c107ceacee7"} Dec 03 00:10:52 crc kubenswrapper[4953]: I1203 00:10:52.508297 4953 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="93b066fdbbc97d1257f39b14356258c5c35804fe30647b77af9b6c107ceacee7" Dec 03 00:10:52 crc kubenswrapper[4953]: I1203 00:10:52.508125 4953 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 03 00:11:03 crc kubenswrapper[4953]: I1203 00:11:03.577625 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-js4dj" event={"ID":"f17dcaea-cb22-4359-9480-e1d129cd309e","Type":"ContainerStarted","Data":"f49059b90bc3bb72cd6d33fb965254b786fdd5681a910b3d6a6cd6cdb7a67582"} Dec 03 00:11:03 crc kubenswrapper[4953]: I1203 00:11:03.580214 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-q94c6" event={"ID":"be2fa7e6-200e-44f0-9b40-57ed2d30db95","Type":"ContainerStarted","Data":"5f47afabcdf2fcef48d1f47b76c2b7afd9f9be5addc6467757669d8b92a67cdc"} Dec 03 00:11:04 crc kubenswrapper[4953]: I1203 00:11:04.590882 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xfd2l" event={"ID":"52fd5773-3888-49f1-813d-10d262f4c901","Type":"ContainerStarted","Data":"3a3bd25c8318f951154af34ed59bdf755743e6f4b6d7bc248b6828f0b46d7888"} Dec 03 00:11:04 crc kubenswrapper[4953]: I1203 00:11:04.592912 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ct49l" event={"ID":"3e5a99ac-9abf-4291-b773-fda0c4364959","Type":"ContainerStarted","Data":"f7a224f432762e714a147eab9d41df6a20e7380726428560f8a488359427d934"} Dec 03 00:11:04 crc kubenswrapper[4953]: I1203 00:11:04.594616 4953 generic.go:334] "Generic (PLEG): container finished" podID="496add53-6511-4560-82c5-3efeb65935f0" containerID="8ca5e10d1bed3228a3d19a1f7415a0d5902d680e20d9f6f913f9501e9bb74a8c" exitCode=0 Dec 03 00:11:04 crc kubenswrapper[4953]: I1203 00:11:04.594682 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6kz7m" event={"ID":"496add53-6511-4560-82c5-3efeb65935f0","Type":"ContainerDied","Data":"8ca5e10d1bed3228a3d19a1f7415a0d5902d680e20d9f6f913f9501e9bb74a8c"} Dec 03 00:11:04 crc kubenswrapper[4953]: I1203 00:11:04.598205 4953 generic.go:334] "Generic (PLEG): container finished" podID="f17dcaea-cb22-4359-9480-e1d129cd309e" containerID="f49059b90bc3bb72cd6d33fb965254b786fdd5681a910b3d6a6cd6cdb7a67582" exitCode=0 Dec 03 00:11:04 crc kubenswrapper[4953]: I1203 00:11:04.598291 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-js4dj" event={"ID":"f17dcaea-cb22-4359-9480-e1d129cd309e","Type":"ContainerDied","Data":"f49059b90bc3bb72cd6d33fb965254b786fdd5681a910b3d6a6cd6cdb7a67582"} Dec 03 00:11:04 crc kubenswrapper[4953]: I1203 00:11:04.604539 4953 generic.go:334] "Generic (PLEG): container finished" podID="be2fa7e6-200e-44f0-9b40-57ed2d30db95" containerID="5f47afabcdf2fcef48d1f47b76c2b7afd9f9be5addc6467757669d8b92a67cdc" exitCode=0 Dec 03 00:11:04 crc kubenswrapper[4953]: I1203 00:11:04.604608 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-q94c6" event={"ID":"be2fa7e6-200e-44f0-9b40-57ed2d30db95","Type":"ContainerDied","Data":"5f47afabcdf2fcef48d1f47b76c2b7afd9f9be5addc6467757669d8b92a67cdc"} Dec 03 00:11:04 crc kubenswrapper[4953]: I1203 00:11:04.617947 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dq6nn" event={"ID":"9064a5ab-24ee-4258-8363-6ce4ffa54b3f","Type":"ContainerStarted","Data":"5409b03d52f75ac22e98a311685907066f3577a6782501f328e36f36f67641c7"} Dec 03 00:11:04 crc kubenswrapper[4953]: I1203 00:11:04.630313 4953 generic.go:334] "Generic (PLEG): container finished" podID="e98b8740-f6fc-4db0-a807-664cf085391c" containerID="ab8f2200b1995da972fc62acce6b9433d30b466901c9ab10b23bbb9a4e2c4bf4" exitCode=0 Dec 03 00:11:04 crc kubenswrapper[4953]: I1203 00:11:04.630372 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5zxsj" event={"ID":"e98b8740-f6fc-4db0-a807-664cf085391c","Type":"ContainerDied","Data":"ab8f2200b1995da972fc62acce6b9433d30b466901c9ab10b23bbb9a4e2c4bf4"} Dec 03 00:11:05 crc kubenswrapper[4953]: I1203 00:11:05.638943 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5zxsj" event={"ID":"e98b8740-f6fc-4db0-a807-664cf085391c","Type":"ContainerStarted","Data":"2bfe87b69385d5610e7e2c6550a4140ae496768d5035f606a6264aad12cc44e5"} Dec 03 00:11:05 crc kubenswrapper[4953]: I1203 00:11:05.642881 4953 generic.go:334] "Generic (PLEG): container finished" podID="52fd5773-3888-49f1-813d-10d262f4c901" containerID="3a3bd25c8318f951154af34ed59bdf755743e6f4b6d7bc248b6828f0b46d7888" exitCode=0 Dec 03 00:11:05 crc kubenswrapper[4953]: I1203 00:11:05.642944 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xfd2l" event={"ID":"52fd5773-3888-49f1-813d-10d262f4c901","Type":"ContainerDied","Data":"3a3bd25c8318f951154af34ed59bdf755743e6f4b6d7bc248b6828f0b46d7888"} Dec 03 00:11:05 crc kubenswrapper[4953]: I1203 00:11:05.647198 4953 generic.go:334] "Generic (PLEG): container finished" podID="3e5a99ac-9abf-4291-b773-fda0c4364959" containerID="f7a224f432762e714a147eab9d41df6a20e7380726428560f8a488359427d934" exitCode=0 Dec 03 00:11:05 crc kubenswrapper[4953]: I1203 00:11:05.647377 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ct49l" event={"ID":"3e5a99ac-9abf-4291-b773-fda0c4364959","Type":"ContainerDied","Data":"f7a224f432762e714a147eab9d41df6a20e7380726428560f8a488359427d934"} Dec 03 00:11:05 crc kubenswrapper[4953]: I1203 00:11:05.652538 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6kz7m" event={"ID":"496add53-6511-4560-82c5-3efeb65935f0","Type":"ContainerStarted","Data":"845879064ed02c31d83bd0919ef8a32ffc34fbe837fe413c5707f2a5bc481852"} Dec 03 00:11:05 crc kubenswrapper[4953]: I1203 00:11:05.655344 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-js4dj" event={"ID":"f17dcaea-cb22-4359-9480-e1d129cd309e","Type":"ContainerStarted","Data":"abff23986f846102f531c304d4701682b0d7a1eacc0e50a4a5efe2534394270f"} Dec 03 00:11:05 crc kubenswrapper[4953]: I1203 00:11:05.665430 4953 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-5zxsj" podStartSLOduration=5.416466198 podStartE2EDuration="1m34.66540101s" podCreationTimestamp="2025-12-03 00:09:31 +0000 UTC" firstStartedPulling="2025-12-03 00:09:35.84737316 +0000 UTC m=+164.831653953" lastFinishedPulling="2025-12-03 00:11:05.096307982 +0000 UTC m=+254.080588765" observedRunningTime="2025-12-03 00:11:05.661238257 +0000 UTC m=+254.645519040" watchObservedRunningTime="2025-12-03 00:11:05.66540101 +0000 UTC m=+254.649681813" Dec 03 00:11:05 crc kubenswrapper[4953]: I1203 00:11:05.669474 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-q94c6" event={"ID":"be2fa7e6-200e-44f0-9b40-57ed2d30db95","Type":"ContainerStarted","Data":"c8db9401dd0016729bfd4ca881af6640ca46a36c96bd34a067b3948e32990743"} Dec 03 00:11:05 crc kubenswrapper[4953]: I1203 00:11:05.678186 4953 generic.go:334] "Generic (PLEG): container finished" podID="9064a5ab-24ee-4258-8363-6ce4ffa54b3f" containerID="5409b03d52f75ac22e98a311685907066f3577a6782501f328e36f36f67641c7" exitCode=0 Dec 03 00:11:05 crc kubenswrapper[4953]: I1203 00:11:05.678300 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dq6nn" event={"ID":"9064a5ab-24ee-4258-8363-6ce4ffa54b3f","Type":"ContainerDied","Data":"5409b03d52f75ac22e98a311685907066f3577a6782501f328e36f36f67641c7"} Dec 03 00:11:05 crc kubenswrapper[4953]: I1203 00:11:05.695060 4953 generic.go:334] "Generic (PLEG): container finished" podID="6b353eef-b7f9-4dbc-b2aa-9860220438cc" containerID="b47d72f0f01ce88a842fca13012bced12a6ffb7e74cf75e86142b1f1c4300e80" exitCode=0 Dec 03 00:11:05 crc kubenswrapper[4953]: I1203 00:11:05.695148 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hs6pg" event={"ID":"6b353eef-b7f9-4dbc-b2aa-9860220438cc","Type":"ContainerDied","Data":"b47d72f0f01ce88a842fca13012bced12a6ffb7e74cf75e86142b1f1c4300e80"} Dec 03 00:11:05 crc kubenswrapper[4953]: I1203 00:11:05.737934 4953 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-js4dj" podStartSLOduration=6.572615087 podStartE2EDuration="1m35.737908916s" podCreationTimestamp="2025-12-03 00:09:30 +0000 UTC" firstStartedPulling="2025-12-03 00:09:35.826274273 +0000 UTC m=+164.810555076" lastFinishedPulling="2025-12-03 00:11:04.991568122 +0000 UTC m=+253.975848905" observedRunningTime="2025-12-03 00:11:05.7340164 +0000 UTC m=+254.718297193" watchObservedRunningTime="2025-12-03 00:11:05.737908916 +0000 UTC m=+254.722189699" Dec 03 00:11:05 crc kubenswrapper[4953]: I1203 00:11:05.837248 4953 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-q94c6" podStartSLOduration=3.581808443 podStartE2EDuration="1m32.837222342s" podCreationTimestamp="2025-12-03 00:09:33 +0000 UTC" firstStartedPulling="2025-12-03 00:09:35.82818008 +0000 UTC m=+164.812460903" lastFinishedPulling="2025-12-03 00:11:05.083594019 +0000 UTC m=+254.067874802" observedRunningTime="2025-12-03 00:11:05.836809952 +0000 UTC m=+254.821090755" watchObservedRunningTime="2025-12-03 00:11:05.837222342 +0000 UTC m=+254.821503125" Dec 03 00:11:06 crc kubenswrapper[4953]: I1203 00:11:06.716625 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xfd2l" event={"ID":"52fd5773-3888-49f1-813d-10d262f4c901","Type":"ContainerStarted","Data":"07abd4b77b78878cb3704fdc549bbd9b8f175c5bb93a78c71f5b651ea9c6e2f8"} Dec 03 00:11:06 crc kubenswrapper[4953]: I1203 00:11:06.719204 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ct49l" event={"ID":"3e5a99ac-9abf-4291-b773-fda0c4364959","Type":"ContainerStarted","Data":"48186f55e61fb9098b010d3501a137fe66f3982d1dbf4701955d4bbe56b52766"} Dec 03 00:11:06 crc kubenswrapper[4953]: I1203 00:11:06.721259 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dq6nn" event={"ID":"9064a5ab-24ee-4258-8363-6ce4ffa54b3f","Type":"ContainerStarted","Data":"d585b7d4d02d5b26ee4092ba449d6de3aa3d6e10bf9bc47f915048d6497ea65e"} Dec 03 00:11:06 crc kubenswrapper[4953]: I1203 00:11:06.726271 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hs6pg" event={"ID":"6b353eef-b7f9-4dbc-b2aa-9860220438cc","Type":"ContainerStarted","Data":"8af485ffdb0db0916a7f4f533b0b05a77458193f57ac4514e40e6bfc8205532f"} Dec 03 00:11:06 crc kubenswrapper[4953]: I1203 00:11:06.742621 4953 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-xfd2l" podStartSLOduration=6.266875741 podStartE2EDuration="1m36.742606201s" podCreationTimestamp="2025-12-03 00:09:30 +0000 UTC" firstStartedPulling="2025-12-03 00:09:35.850940587 +0000 UTC m=+164.835221380" lastFinishedPulling="2025-12-03 00:11:06.326671057 +0000 UTC m=+255.310951840" observedRunningTime="2025-12-03 00:11:06.739376412 +0000 UTC m=+255.723657195" watchObservedRunningTime="2025-12-03 00:11:06.742606201 +0000 UTC m=+255.726886984" Dec 03 00:11:06 crc kubenswrapper[4953]: I1203 00:11:06.763485 4953 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-6kz7m" podStartSLOduration=5.387625065 podStartE2EDuration="1m34.763468105s" podCreationTimestamp="2025-12-03 00:09:32 +0000 UTC" firstStartedPulling="2025-12-03 00:09:35.868912958 +0000 UTC m=+164.853193751" lastFinishedPulling="2025-12-03 00:11:05.244756008 +0000 UTC m=+254.229036791" observedRunningTime="2025-12-03 00:11:06.761741043 +0000 UTC m=+255.746021826" watchObservedRunningTime="2025-12-03 00:11:06.763468105 +0000 UTC m=+255.747748888" Dec 03 00:11:06 crc kubenswrapper[4953]: I1203 00:11:06.789682 4953 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-dq6nn" podStartSLOduration=3.319369358 podStartE2EDuration="1m33.789666721s" podCreationTimestamp="2025-12-03 00:09:33 +0000 UTC" firstStartedPulling="2025-12-03 00:09:35.830958108 +0000 UTC m=+164.815238901" lastFinishedPulling="2025-12-03 00:11:06.301255481 +0000 UTC m=+255.285536264" observedRunningTime="2025-12-03 00:11:06.787111908 +0000 UTC m=+255.771392711" watchObservedRunningTime="2025-12-03 00:11:06.789666721 +0000 UTC m=+255.773947504" Dec 03 00:11:06 crc kubenswrapper[4953]: I1203 00:11:06.844535 4953 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-ct49l" podStartSLOduration=6.5168760169999995 podStartE2EDuration="1m36.844502031s" podCreationTimestamp="2025-12-03 00:09:30 +0000 UTC" firstStartedPulling="2025-12-03 00:09:35.875486069 +0000 UTC m=+164.859766892" lastFinishedPulling="2025-12-03 00:11:06.203112123 +0000 UTC m=+255.187392906" observedRunningTime="2025-12-03 00:11:06.844273006 +0000 UTC m=+255.828553789" watchObservedRunningTime="2025-12-03 00:11:06.844502031 +0000 UTC m=+255.828782814" Dec 03 00:11:06 crc kubenswrapper[4953]: I1203 00:11:06.847573 4953 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-hs6pg" podStartSLOduration=6.397873101 podStartE2EDuration="1m36.847562807s" podCreationTimestamp="2025-12-03 00:09:30 +0000 UTC" firstStartedPulling="2025-12-03 00:09:35.834772951 +0000 UTC m=+164.819053754" lastFinishedPulling="2025-12-03 00:11:06.284462677 +0000 UTC m=+255.268743460" observedRunningTime="2025-12-03 00:11:06.813332934 +0000 UTC m=+255.797613717" watchObservedRunningTime="2025-12-03 00:11:06.847562807 +0000 UTC m=+255.831843580" Dec 03 00:11:10 crc kubenswrapper[4953]: I1203 00:11:10.444568 4953 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-hs6pg" Dec 03 00:11:10 crc kubenswrapper[4953]: I1203 00:11:10.445167 4953 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-hs6pg" Dec 03 00:11:10 crc kubenswrapper[4953]: I1203 00:11:10.469255 4953 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-xfd2l" Dec 03 00:11:10 crc kubenswrapper[4953]: I1203 00:11:10.469357 4953 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-xfd2l" Dec 03 00:11:10 crc kubenswrapper[4953]: I1203 00:11:10.535624 4953 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-hs6pg" Dec 03 00:11:10 crc kubenswrapper[4953]: I1203 00:11:10.545470 4953 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-xfd2l" Dec 03 00:11:10 crc kubenswrapper[4953]: I1203 00:11:10.605729 4953 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-ct49l" Dec 03 00:11:10 crc kubenswrapper[4953]: I1203 00:11:10.605796 4953 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-ct49l" Dec 03 00:11:10 crc kubenswrapper[4953]: I1203 00:11:10.654539 4953 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-ct49l" Dec 03 00:11:10 crc kubenswrapper[4953]: I1203 00:11:10.905431 4953 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-js4dj" Dec 03 00:11:10 crc kubenswrapper[4953]: I1203 00:11:10.905765 4953 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-js4dj" Dec 03 00:11:10 crc kubenswrapper[4953]: I1203 00:11:10.969965 4953 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-js4dj" Dec 03 00:11:11 crc kubenswrapper[4953]: I1203 00:11:11.812929 4953 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-js4dj" Dec 03 00:11:12 crc kubenswrapper[4953]: I1203 00:11:12.160926 4953 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-5zxsj" Dec 03 00:11:12 crc kubenswrapper[4953]: I1203 00:11:12.161013 4953 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-5zxsj" Dec 03 00:11:12 crc kubenswrapper[4953]: I1203 00:11:12.327579 4953 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-5zxsj" Dec 03 00:11:12 crc kubenswrapper[4953]: I1203 00:11:12.632028 4953 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-6kz7m" Dec 03 00:11:12 crc kubenswrapper[4953]: I1203 00:11:12.632110 4953 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-6kz7m" Dec 03 00:11:12 crc kubenswrapper[4953]: I1203 00:11:12.702436 4953 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-6kz7m" Dec 03 00:11:12 crc kubenswrapper[4953]: I1203 00:11:12.809840 4953 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-5zxsj" Dec 03 00:11:12 crc kubenswrapper[4953]: I1203 00:11:12.814547 4953 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-6kz7m" Dec 03 00:11:13 crc kubenswrapper[4953]: I1203 00:11:13.593180 4953 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-js4dj"] Dec 03 00:11:13 crc kubenswrapper[4953]: I1203 00:11:13.611267 4953 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-dq6nn" Dec 03 00:11:13 crc kubenswrapper[4953]: I1203 00:11:13.611360 4953 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-dq6nn" Dec 03 00:11:13 crc kubenswrapper[4953]: I1203 00:11:13.669515 4953 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-dq6nn" Dec 03 00:11:13 crc kubenswrapper[4953]: I1203 00:11:13.766741 4953 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-js4dj" podUID="f17dcaea-cb22-4359-9480-e1d129cd309e" containerName="registry-server" containerID="cri-o://abff23986f846102f531c304d4701682b0d7a1eacc0e50a4a5efe2534394270f" gracePeriod=2 Dec 03 00:11:13 crc kubenswrapper[4953]: I1203 00:11:13.814631 4953 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-dq6nn" Dec 03 00:11:13 crc kubenswrapper[4953]: I1203 00:11:13.975357 4953 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-q94c6" Dec 03 00:11:13 crc kubenswrapper[4953]: I1203 00:11:13.975411 4953 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-q94c6" Dec 03 00:11:14 crc kubenswrapper[4953]: I1203 00:11:14.013151 4953 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-q94c6" Dec 03 00:11:14 crc kubenswrapper[4953]: I1203 00:11:14.832382 4953 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-q94c6" Dec 03 00:11:15 crc kubenswrapper[4953]: I1203 00:11:15.784446 4953 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-6kz7m"] Dec 03 00:11:15 crc kubenswrapper[4953]: I1203 00:11:15.784802 4953 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-6kz7m" podUID="496add53-6511-4560-82c5-3efeb65935f0" containerName="registry-server" containerID="cri-o://845879064ed02c31d83bd0919ef8a32ffc34fbe837fe413c5707f2a5bc481852" gracePeriod=2 Dec 03 00:11:16 crc kubenswrapper[4953]: I1203 00:11:16.792362 4953 generic.go:334] "Generic (PLEG): container finished" podID="496add53-6511-4560-82c5-3efeb65935f0" containerID="845879064ed02c31d83bd0919ef8a32ffc34fbe837fe413c5707f2a5bc481852" exitCode=0 Dec 03 00:11:16 crc kubenswrapper[4953]: I1203 00:11:16.792483 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6kz7m" event={"ID":"496add53-6511-4560-82c5-3efeb65935f0","Type":"ContainerDied","Data":"845879064ed02c31d83bd0919ef8a32ffc34fbe837fe413c5707f2a5bc481852"} Dec 03 00:11:16 crc kubenswrapper[4953]: I1203 00:11:16.795609 4953 generic.go:334] "Generic (PLEG): container finished" podID="f17dcaea-cb22-4359-9480-e1d129cd309e" containerID="abff23986f846102f531c304d4701682b0d7a1eacc0e50a4a5efe2534394270f" exitCode=0 Dec 03 00:11:16 crc kubenswrapper[4953]: I1203 00:11:16.795663 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-js4dj" event={"ID":"f17dcaea-cb22-4359-9480-e1d129cd309e","Type":"ContainerDied","Data":"abff23986f846102f531c304d4701682b0d7a1eacc0e50a4a5efe2534394270f"} Dec 03 00:11:17 crc kubenswrapper[4953]: I1203 00:11:17.198874 4953 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-js4dj" Dec 03 00:11:17 crc kubenswrapper[4953]: I1203 00:11:17.338459 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f17dcaea-cb22-4359-9480-e1d129cd309e-utilities\") pod \"f17dcaea-cb22-4359-9480-e1d129cd309e\" (UID: \"f17dcaea-cb22-4359-9480-e1d129cd309e\") " Dec 03 00:11:17 crc kubenswrapper[4953]: I1203 00:11:17.338581 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jfmxh\" (UniqueName: \"kubernetes.io/projected/f17dcaea-cb22-4359-9480-e1d129cd309e-kube-api-access-jfmxh\") pod \"f17dcaea-cb22-4359-9480-e1d129cd309e\" (UID: \"f17dcaea-cb22-4359-9480-e1d129cd309e\") " Dec 03 00:11:17 crc kubenswrapper[4953]: I1203 00:11:17.338657 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f17dcaea-cb22-4359-9480-e1d129cd309e-catalog-content\") pod \"f17dcaea-cb22-4359-9480-e1d129cd309e\" (UID: \"f17dcaea-cb22-4359-9480-e1d129cd309e\") " Dec 03 00:11:17 crc kubenswrapper[4953]: I1203 00:11:17.339865 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f17dcaea-cb22-4359-9480-e1d129cd309e-utilities" (OuterVolumeSpecName: "utilities") pod "f17dcaea-cb22-4359-9480-e1d129cd309e" (UID: "f17dcaea-cb22-4359-9480-e1d129cd309e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 00:11:17 crc kubenswrapper[4953]: I1203 00:11:17.348046 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f17dcaea-cb22-4359-9480-e1d129cd309e-kube-api-access-jfmxh" (OuterVolumeSpecName: "kube-api-access-jfmxh") pod "f17dcaea-cb22-4359-9480-e1d129cd309e" (UID: "f17dcaea-cb22-4359-9480-e1d129cd309e"). InnerVolumeSpecName "kube-api-access-jfmxh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:11:17 crc kubenswrapper[4953]: I1203 00:11:17.440509 4953 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f17dcaea-cb22-4359-9480-e1d129cd309e-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 00:11:17 crc kubenswrapper[4953]: I1203 00:11:17.440555 4953 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jfmxh\" (UniqueName: \"kubernetes.io/projected/f17dcaea-cb22-4359-9480-e1d129cd309e-kube-api-access-jfmxh\") on node \"crc\" DevicePath \"\"" Dec 03 00:11:17 crc kubenswrapper[4953]: I1203 00:11:17.612633 4953 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-6kz7m" Dec 03 00:11:17 crc kubenswrapper[4953]: I1203 00:11:17.744452 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/496add53-6511-4560-82c5-3efeb65935f0-utilities\") pod \"496add53-6511-4560-82c5-3efeb65935f0\" (UID: \"496add53-6511-4560-82c5-3efeb65935f0\") " Dec 03 00:11:17 crc kubenswrapper[4953]: I1203 00:11:17.744515 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/496add53-6511-4560-82c5-3efeb65935f0-catalog-content\") pod \"496add53-6511-4560-82c5-3efeb65935f0\" (UID: \"496add53-6511-4560-82c5-3efeb65935f0\") " Dec 03 00:11:17 crc kubenswrapper[4953]: I1203 00:11:17.744732 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gjvz8\" (UniqueName: \"kubernetes.io/projected/496add53-6511-4560-82c5-3efeb65935f0-kube-api-access-gjvz8\") pod \"496add53-6511-4560-82c5-3efeb65935f0\" (UID: \"496add53-6511-4560-82c5-3efeb65935f0\") " Dec 03 00:11:17 crc kubenswrapper[4953]: I1203 00:11:17.745588 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/496add53-6511-4560-82c5-3efeb65935f0-utilities" (OuterVolumeSpecName: "utilities") pod "496add53-6511-4560-82c5-3efeb65935f0" (UID: "496add53-6511-4560-82c5-3efeb65935f0"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 00:11:17 crc kubenswrapper[4953]: I1203 00:11:17.749421 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/496add53-6511-4560-82c5-3efeb65935f0-kube-api-access-gjvz8" (OuterVolumeSpecName: "kube-api-access-gjvz8") pod "496add53-6511-4560-82c5-3efeb65935f0" (UID: "496add53-6511-4560-82c5-3efeb65935f0"). InnerVolumeSpecName "kube-api-access-gjvz8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:11:17 crc kubenswrapper[4953]: I1203 00:11:17.766535 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/496add53-6511-4560-82c5-3efeb65935f0-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "496add53-6511-4560-82c5-3efeb65935f0" (UID: "496add53-6511-4560-82c5-3efeb65935f0"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 00:11:17 crc kubenswrapper[4953]: I1203 00:11:17.806258 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-js4dj" event={"ID":"f17dcaea-cb22-4359-9480-e1d129cd309e","Type":"ContainerDied","Data":"00378a6caa696020038368823cb33ae715cae36617efa82c26df2c5c5d2708e6"} Dec 03 00:11:17 crc kubenswrapper[4953]: I1203 00:11:17.806364 4953 scope.go:117] "RemoveContainer" containerID="abff23986f846102f531c304d4701682b0d7a1eacc0e50a4a5efe2534394270f" Dec 03 00:11:17 crc kubenswrapper[4953]: I1203 00:11:17.806669 4953 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-js4dj" Dec 03 00:11:17 crc kubenswrapper[4953]: I1203 00:11:17.812191 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6kz7m" event={"ID":"496add53-6511-4560-82c5-3efeb65935f0","Type":"ContainerDied","Data":"eadf7e7493d10cc4697afbeac29011f04e4fa41b0dc506c07df8179ab4703e75"} Dec 03 00:11:17 crc kubenswrapper[4953]: I1203 00:11:17.812323 4953 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-6kz7m" Dec 03 00:11:17 crc kubenswrapper[4953]: I1203 00:11:17.827814 4953 scope.go:117] "RemoveContainer" containerID="f49059b90bc3bb72cd6d33fb965254b786fdd5681a910b3d6a6cd6cdb7a67582" Dec 03 00:11:17 crc kubenswrapper[4953]: I1203 00:11:17.847470 4953 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/496add53-6511-4560-82c5-3efeb65935f0-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 00:11:17 crc kubenswrapper[4953]: I1203 00:11:17.854306 4953 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/496add53-6511-4560-82c5-3efeb65935f0-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 00:11:17 crc kubenswrapper[4953]: I1203 00:11:17.854446 4953 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gjvz8\" (UniqueName: \"kubernetes.io/projected/496add53-6511-4560-82c5-3efeb65935f0-kube-api-access-gjvz8\") on node \"crc\" DevicePath \"\"" Dec 03 00:11:17 crc kubenswrapper[4953]: I1203 00:11:17.861769 4953 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-6kz7m"] Dec 03 00:11:17 crc kubenswrapper[4953]: I1203 00:11:17.868340 4953 scope.go:117] "RemoveContainer" containerID="2514370e8db161da2ec1d92134a419afbf628145aa9c9a74b56340bc8b6ce59e" Dec 03 00:11:17 crc kubenswrapper[4953]: I1203 00:11:17.869842 4953 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-6kz7m"] Dec 03 00:11:17 crc kubenswrapper[4953]: I1203 00:11:17.901441 4953 scope.go:117] "RemoveContainer" containerID="845879064ed02c31d83bd0919ef8a32ffc34fbe837fe413c5707f2a5bc481852" Dec 03 00:11:17 crc kubenswrapper[4953]: I1203 00:11:17.919495 4953 scope.go:117] "RemoveContainer" containerID="8ca5e10d1bed3228a3d19a1f7415a0d5902d680e20d9f6f913f9501e9bb74a8c" Dec 03 00:11:17 crc kubenswrapper[4953]: I1203 00:11:17.934352 4953 scope.go:117] "RemoveContainer" containerID="57759415d3b5338ee044069760976184c643ecf174ff7fdd85eea90738b2fdea" Dec 03 00:11:18 crc kubenswrapper[4953]: I1203 00:11:18.184782 4953 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-q94c6"] Dec 03 00:11:18 crc kubenswrapper[4953]: I1203 00:11:18.185589 4953 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-q94c6" podUID="be2fa7e6-200e-44f0-9b40-57ed2d30db95" containerName="registry-server" containerID="cri-o://c8db9401dd0016729bfd4ca881af6640ca46a36c96bd34a067b3948e32990743" gracePeriod=2 Dec 03 00:11:18 crc kubenswrapper[4953]: I1203 00:11:18.453749 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f17dcaea-cb22-4359-9480-e1d129cd309e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f17dcaea-cb22-4359-9480-e1d129cd309e" (UID: "f17dcaea-cb22-4359-9480-e1d129cd309e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 00:11:18 crc kubenswrapper[4953]: I1203 00:11:18.463684 4953 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f17dcaea-cb22-4359-9480-e1d129cd309e-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 00:11:18 crc kubenswrapper[4953]: I1203 00:11:18.746934 4953 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-js4dj"] Dec 03 00:11:18 crc kubenswrapper[4953]: I1203 00:11:18.751495 4953 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-js4dj"] Dec 03 00:11:18 crc kubenswrapper[4953]: I1203 00:11:18.856938 4953 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-45k44"] Dec 03 00:11:19 crc kubenswrapper[4953]: I1203 00:11:19.357655 4953 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="496add53-6511-4560-82c5-3efeb65935f0" path="/var/lib/kubelet/pods/496add53-6511-4560-82c5-3efeb65935f0/volumes" Dec 03 00:11:19 crc kubenswrapper[4953]: I1203 00:11:19.359025 4953 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f17dcaea-cb22-4359-9480-e1d129cd309e" path="/var/lib/kubelet/pods/f17dcaea-cb22-4359-9480-e1d129cd309e/volumes" Dec 03 00:11:19 crc kubenswrapper[4953]: I1203 00:11:19.441297 4953 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-q94c6" Dec 03 00:11:19 crc kubenswrapper[4953]: I1203 00:11:19.478457 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ghrbv\" (UniqueName: \"kubernetes.io/projected/be2fa7e6-200e-44f0-9b40-57ed2d30db95-kube-api-access-ghrbv\") pod \"be2fa7e6-200e-44f0-9b40-57ed2d30db95\" (UID: \"be2fa7e6-200e-44f0-9b40-57ed2d30db95\") " Dec 03 00:11:19 crc kubenswrapper[4953]: I1203 00:11:19.478619 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/be2fa7e6-200e-44f0-9b40-57ed2d30db95-catalog-content\") pod \"be2fa7e6-200e-44f0-9b40-57ed2d30db95\" (UID: \"be2fa7e6-200e-44f0-9b40-57ed2d30db95\") " Dec 03 00:11:19 crc kubenswrapper[4953]: I1203 00:11:19.478927 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/be2fa7e6-200e-44f0-9b40-57ed2d30db95-utilities\") pod \"be2fa7e6-200e-44f0-9b40-57ed2d30db95\" (UID: \"be2fa7e6-200e-44f0-9b40-57ed2d30db95\") " Dec 03 00:11:19 crc kubenswrapper[4953]: I1203 00:11:19.480146 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/be2fa7e6-200e-44f0-9b40-57ed2d30db95-utilities" (OuterVolumeSpecName: "utilities") pod "be2fa7e6-200e-44f0-9b40-57ed2d30db95" (UID: "be2fa7e6-200e-44f0-9b40-57ed2d30db95"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 00:11:19 crc kubenswrapper[4953]: I1203 00:11:19.485081 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/be2fa7e6-200e-44f0-9b40-57ed2d30db95-kube-api-access-ghrbv" (OuterVolumeSpecName: "kube-api-access-ghrbv") pod "be2fa7e6-200e-44f0-9b40-57ed2d30db95" (UID: "be2fa7e6-200e-44f0-9b40-57ed2d30db95"). InnerVolumeSpecName "kube-api-access-ghrbv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:11:19 crc kubenswrapper[4953]: I1203 00:11:19.581513 4953 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ghrbv\" (UniqueName: \"kubernetes.io/projected/be2fa7e6-200e-44f0-9b40-57ed2d30db95-kube-api-access-ghrbv\") on node \"crc\" DevicePath \"\"" Dec 03 00:11:19 crc kubenswrapper[4953]: I1203 00:11:19.581559 4953 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/be2fa7e6-200e-44f0-9b40-57ed2d30db95-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 00:11:19 crc kubenswrapper[4953]: I1203 00:11:19.830936 4953 generic.go:334] "Generic (PLEG): container finished" podID="be2fa7e6-200e-44f0-9b40-57ed2d30db95" containerID="c8db9401dd0016729bfd4ca881af6640ca46a36c96bd34a067b3948e32990743" exitCode=0 Dec 03 00:11:19 crc kubenswrapper[4953]: I1203 00:11:19.831040 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-q94c6" event={"ID":"be2fa7e6-200e-44f0-9b40-57ed2d30db95","Type":"ContainerDied","Data":"c8db9401dd0016729bfd4ca881af6640ca46a36c96bd34a067b3948e32990743"} Dec 03 00:11:19 crc kubenswrapper[4953]: I1203 00:11:19.831081 4953 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-q94c6" Dec 03 00:11:19 crc kubenswrapper[4953]: I1203 00:11:19.831121 4953 scope.go:117] "RemoveContainer" containerID="c8db9401dd0016729bfd4ca881af6640ca46a36c96bd34a067b3948e32990743" Dec 03 00:11:19 crc kubenswrapper[4953]: I1203 00:11:19.831105 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-q94c6" event={"ID":"be2fa7e6-200e-44f0-9b40-57ed2d30db95","Type":"ContainerDied","Data":"c64ff0759b4e0582da669df0d162803598a061dd7c8baa5aef06f79aba6470ac"} Dec 03 00:11:19 crc kubenswrapper[4953]: I1203 00:11:19.850558 4953 scope.go:117] "RemoveContainer" containerID="5f47afabcdf2fcef48d1f47b76c2b7afd9f9be5addc6467757669d8b92a67cdc" Dec 03 00:11:19 crc kubenswrapper[4953]: I1203 00:11:19.867932 4953 scope.go:117] "RemoveContainer" containerID="7933636e09868e01e20fa43288d4411f1cfd94ad47ec9d021d115adc318c4885" Dec 03 00:11:19 crc kubenswrapper[4953]: I1203 00:11:19.886158 4953 scope.go:117] "RemoveContainer" containerID="c8db9401dd0016729bfd4ca881af6640ca46a36c96bd34a067b3948e32990743" Dec 03 00:11:19 crc kubenswrapper[4953]: E1203 00:11:19.886720 4953 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c8db9401dd0016729bfd4ca881af6640ca46a36c96bd34a067b3948e32990743\": container with ID starting with c8db9401dd0016729bfd4ca881af6640ca46a36c96bd34a067b3948e32990743 not found: ID does not exist" containerID="c8db9401dd0016729bfd4ca881af6640ca46a36c96bd34a067b3948e32990743" Dec 03 00:11:19 crc kubenswrapper[4953]: I1203 00:11:19.886755 4953 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c8db9401dd0016729bfd4ca881af6640ca46a36c96bd34a067b3948e32990743"} err="failed to get container status \"c8db9401dd0016729bfd4ca881af6640ca46a36c96bd34a067b3948e32990743\": rpc error: code = NotFound desc = could not find container \"c8db9401dd0016729bfd4ca881af6640ca46a36c96bd34a067b3948e32990743\": container with ID starting with c8db9401dd0016729bfd4ca881af6640ca46a36c96bd34a067b3948e32990743 not found: ID does not exist" Dec 03 00:11:19 crc kubenswrapper[4953]: I1203 00:11:19.886777 4953 scope.go:117] "RemoveContainer" containerID="5f47afabcdf2fcef48d1f47b76c2b7afd9f9be5addc6467757669d8b92a67cdc" Dec 03 00:11:19 crc kubenswrapper[4953]: E1203 00:11:19.887039 4953 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5f47afabcdf2fcef48d1f47b76c2b7afd9f9be5addc6467757669d8b92a67cdc\": container with ID starting with 5f47afabcdf2fcef48d1f47b76c2b7afd9f9be5addc6467757669d8b92a67cdc not found: ID does not exist" containerID="5f47afabcdf2fcef48d1f47b76c2b7afd9f9be5addc6467757669d8b92a67cdc" Dec 03 00:11:19 crc kubenswrapper[4953]: I1203 00:11:19.887075 4953 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5f47afabcdf2fcef48d1f47b76c2b7afd9f9be5addc6467757669d8b92a67cdc"} err="failed to get container status \"5f47afabcdf2fcef48d1f47b76c2b7afd9f9be5addc6467757669d8b92a67cdc\": rpc error: code = NotFound desc = could not find container \"5f47afabcdf2fcef48d1f47b76c2b7afd9f9be5addc6467757669d8b92a67cdc\": container with ID starting with 5f47afabcdf2fcef48d1f47b76c2b7afd9f9be5addc6467757669d8b92a67cdc not found: ID does not exist" Dec 03 00:11:19 crc kubenswrapper[4953]: I1203 00:11:19.887098 4953 scope.go:117] "RemoveContainer" containerID="7933636e09868e01e20fa43288d4411f1cfd94ad47ec9d021d115adc318c4885" Dec 03 00:11:19 crc kubenswrapper[4953]: E1203 00:11:19.887358 4953 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7933636e09868e01e20fa43288d4411f1cfd94ad47ec9d021d115adc318c4885\": container with ID starting with 7933636e09868e01e20fa43288d4411f1cfd94ad47ec9d021d115adc318c4885 not found: ID does not exist" containerID="7933636e09868e01e20fa43288d4411f1cfd94ad47ec9d021d115adc318c4885" Dec 03 00:11:19 crc kubenswrapper[4953]: I1203 00:11:19.887383 4953 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7933636e09868e01e20fa43288d4411f1cfd94ad47ec9d021d115adc318c4885"} err="failed to get container status \"7933636e09868e01e20fa43288d4411f1cfd94ad47ec9d021d115adc318c4885\": rpc error: code = NotFound desc = could not find container \"7933636e09868e01e20fa43288d4411f1cfd94ad47ec9d021d115adc318c4885\": container with ID starting with 7933636e09868e01e20fa43288d4411f1cfd94ad47ec9d021d115adc318c4885 not found: ID does not exist" Dec 03 00:11:20 crc kubenswrapper[4953]: I1203 00:11:20.438070 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/be2fa7e6-200e-44f0-9b40-57ed2d30db95-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "be2fa7e6-200e-44f0-9b40-57ed2d30db95" (UID: "be2fa7e6-200e-44f0-9b40-57ed2d30db95"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 00:11:20 crc kubenswrapper[4953]: I1203 00:11:20.493714 4953 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-hs6pg" Dec 03 00:11:20 crc kubenswrapper[4953]: I1203 00:11:20.495547 4953 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/be2fa7e6-200e-44f0-9b40-57ed2d30db95-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 00:11:20 crc kubenswrapper[4953]: I1203 00:11:20.528902 4953 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-xfd2l" Dec 03 00:11:20 crc kubenswrapper[4953]: I1203 00:11:20.653894 4953 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-ct49l" Dec 03 00:11:20 crc kubenswrapper[4953]: I1203 00:11:20.765252 4953 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-q94c6"] Dec 03 00:11:20 crc kubenswrapper[4953]: I1203 00:11:20.768811 4953 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-q94c6"] Dec 03 00:11:21 crc kubenswrapper[4953]: I1203 00:11:21.363439 4953 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="be2fa7e6-200e-44f0-9b40-57ed2d30db95" path="/var/lib/kubelet/pods/be2fa7e6-200e-44f0-9b40-57ed2d30db95/volumes" Dec 03 00:11:24 crc kubenswrapper[4953]: I1203 00:11:24.383814 4953 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-ct49l"] Dec 03 00:11:24 crc kubenswrapper[4953]: I1203 00:11:24.385612 4953 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-ct49l" podUID="3e5a99ac-9abf-4291-b773-fda0c4364959" containerName="registry-server" containerID="cri-o://48186f55e61fb9098b010d3501a137fe66f3982d1dbf4701955d4bbe56b52766" gracePeriod=2 Dec 03 00:11:24 crc kubenswrapper[4953]: I1203 00:11:24.764432 4953 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-ct49l" Dec 03 00:11:24 crc kubenswrapper[4953]: I1203 00:11:24.880062 4953 generic.go:334] "Generic (PLEG): container finished" podID="3e5a99ac-9abf-4291-b773-fda0c4364959" containerID="48186f55e61fb9098b010d3501a137fe66f3982d1dbf4701955d4bbe56b52766" exitCode=0 Dec 03 00:11:24 crc kubenswrapper[4953]: I1203 00:11:24.880144 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ct49l" event={"ID":"3e5a99ac-9abf-4291-b773-fda0c4364959","Type":"ContainerDied","Data":"48186f55e61fb9098b010d3501a137fe66f3982d1dbf4701955d4bbe56b52766"} Dec 03 00:11:24 crc kubenswrapper[4953]: I1203 00:11:24.880189 4953 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-ct49l" Dec 03 00:11:24 crc kubenswrapper[4953]: I1203 00:11:24.880222 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ct49l" event={"ID":"3e5a99ac-9abf-4291-b773-fda0c4364959","Type":"ContainerDied","Data":"78ef4c2de9211ad620e2f2ce4fa3886bea7baa989225de5e229e15ace101e0b7"} Dec 03 00:11:24 crc kubenswrapper[4953]: I1203 00:11:24.880252 4953 scope.go:117] "RemoveContainer" containerID="48186f55e61fb9098b010d3501a137fe66f3982d1dbf4701955d4bbe56b52766" Dec 03 00:11:24 crc kubenswrapper[4953]: I1203 00:11:24.904641 4953 scope.go:117] "RemoveContainer" containerID="f7a224f432762e714a147eab9d41df6a20e7380726428560f8a488359427d934" Dec 03 00:11:24 crc kubenswrapper[4953]: I1203 00:11:24.922083 4953 scope.go:117] "RemoveContainer" containerID="9fbad7d01aef12b68dbe8856ab62492197d7d0237185915678fa26b4609f5e54" Dec 03 00:11:24 crc kubenswrapper[4953]: I1203 00:11:24.943542 4953 scope.go:117] "RemoveContainer" containerID="48186f55e61fb9098b010d3501a137fe66f3982d1dbf4701955d4bbe56b52766" Dec 03 00:11:24 crc kubenswrapper[4953]: E1203 00:11:24.944260 4953 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"48186f55e61fb9098b010d3501a137fe66f3982d1dbf4701955d4bbe56b52766\": container with ID starting with 48186f55e61fb9098b010d3501a137fe66f3982d1dbf4701955d4bbe56b52766 not found: ID does not exist" containerID="48186f55e61fb9098b010d3501a137fe66f3982d1dbf4701955d4bbe56b52766" Dec 03 00:11:24 crc kubenswrapper[4953]: I1203 00:11:24.944365 4953 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"48186f55e61fb9098b010d3501a137fe66f3982d1dbf4701955d4bbe56b52766"} err="failed to get container status \"48186f55e61fb9098b010d3501a137fe66f3982d1dbf4701955d4bbe56b52766\": rpc error: code = NotFound desc = could not find container \"48186f55e61fb9098b010d3501a137fe66f3982d1dbf4701955d4bbe56b52766\": container with ID starting with 48186f55e61fb9098b010d3501a137fe66f3982d1dbf4701955d4bbe56b52766 not found: ID does not exist" Dec 03 00:11:24 crc kubenswrapper[4953]: I1203 00:11:24.944441 4953 scope.go:117] "RemoveContainer" containerID="f7a224f432762e714a147eab9d41df6a20e7380726428560f8a488359427d934" Dec 03 00:11:24 crc kubenswrapper[4953]: E1203 00:11:24.944872 4953 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f7a224f432762e714a147eab9d41df6a20e7380726428560f8a488359427d934\": container with ID starting with f7a224f432762e714a147eab9d41df6a20e7380726428560f8a488359427d934 not found: ID does not exist" containerID="f7a224f432762e714a147eab9d41df6a20e7380726428560f8a488359427d934" Dec 03 00:11:24 crc kubenswrapper[4953]: I1203 00:11:24.944918 4953 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f7a224f432762e714a147eab9d41df6a20e7380726428560f8a488359427d934"} err="failed to get container status \"f7a224f432762e714a147eab9d41df6a20e7380726428560f8a488359427d934\": rpc error: code = NotFound desc = could not find container \"f7a224f432762e714a147eab9d41df6a20e7380726428560f8a488359427d934\": container with ID starting with f7a224f432762e714a147eab9d41df6a20e7380726428560f8a488359427d934 not found: ID does not exist" Dec 03 00:11:24 crc kubenswrapper[4953]: I1203 00:11:24.944958 4953 scope.go:117] "RemoveContainer" containerID="9fbad7d01aef12b68dbe8856ab62492197d7d0237185915678fa26b4609f5e54" Dec 03 00:11:24 crc kubenswrapper[4953]: E1203 00:11:24.945248 4953 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9fbad7d01aef12b68dbe8856ab62492197d7d0237185915678fa26b4609f5e54\": container with ID starting with 9fbad7d01aef12b68dbe8856ab62492197d7d0237185915678fa26b4609f5e54 not found: ID does not exist" containerID="9fbad7d01aef12b68dbe8856ab62492197d7d0237185915678fa26b4609f5e54" Dec 03 00:11:24 crc kubenswrapper[4953]: I1203 00:11:24.945274 4953 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9fbad7d01aef12b68dbe8856ab62492197d7d0237185915678fa26b4609f5e54"} err="failed to get container status \"9fbad7d01aef12b68dbe8856ab62492197d7d0237185915678fa26b4609f5e54\": rpc error: code = NotFound desc = could not find container \"9fbad7d01aef12b68dbe8856ab62492197d7d0237185915678fa26b4609f5e54\": container with ID starting with 9fbad7d01aef12b68dbe8856ab62492197d7d0237185915678fa26b4609f5e54 not found: ID does not exist" Dec 03 00:11:24 crc kubenswrapper[4953]: I1203 00:11:24.959935 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3e5a99ac-9abf-4291-b773-fda0c4364959-catalog-content\") pod \"3e5a99ac-9abf-4291-b773-fda0c4364959\" (UID: \"3e5a99ac-9abf-4291-b773-fda0c4364959\") " Dec 03 00:11:24 crc kubenswrapper[4953]: I1203 00:11:24.960132 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lffhw\" (UniqueName: \"kubernetes.io/projected/3e5a99ac-9abf-4291-b773-fda0c4364959-kube-api-access-lffhw\") pod \"3e5a99ac-9abf-4291-b773-fda0c4364959\" (UID: \"3e5a99ac-9abf-4291-b773-fda0c4364959\") " Dec 03 00:11:24 crc kubenswrapper[4953]: I1203 00:11:24.960222 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3e5a99ac-9abf-4291-b773-fda0c4364959-utilities\") pod \"3e5a99ac-9abf-4291-b773-fda0c4364959\" (UID: \"3e5a99ac-9abf-4291-b773-fda0c4364959\") " Dec 03 00:11:24 crc kubenswrapper[4953]: I1203 00:11:24.961653 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3e5a99ac-9abf-4291-b773-fda0c4364959-utilities" (OuterVolumeSpecName: "utilities") pod "3e5a99ac-9abf-4291-b773-fda0c4364959" (UID: "3e5a99ac-9abf-4291-b773-fda0c4364959"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 00:11:24 crc kubenswrapper[4953]: I1203 00:11:24.968442 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3e5a99ac-9abf-4291-b773-fda0c4364959-kube-api-access-lffhw" (OuterVolumeSpecName: "kube-api-access-lffhw") pod "3e5a99ac-9abf-4291-b773-fda0c4364959" (UID: "3e5a99ac-9abf-4291-b773-fda0c4364959"). InnerVolumeSpecName "kube-api-access-lffhw". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:11:25 crc kubenswrapper[4953]: I1203 00:11:25.031195 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3e5a99ac-9abf-4291-b773-fda0c4364959-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "3e5a99ac-9abf-4291-b773-fda0c4364959" (UID: "3e5a99ac-9abf-4291-b773-fda0c4364959"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 00:11:25 crc kubenswrapper[4953]: I1203 00:11:25.063221 4953 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3e5a99ac-9abf-4291-b773-fda0c4364959-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 00:11:25 crc kubenswrapper[4953]: I1203 00:11:25.063286 4953 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3e5a99ac-9abf-4291-b773-fda0c4364959-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 00:11:25 crc kubenswrapper[4953]: I1203 00:11:25.063312 4953 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lffhw\" (UniqueName: \"kubernetes.io/projected/3e5a99ac-9abf-4291-b773-fda0c4364959-kube-api-access-lffhw\") on node \"crc\" DevicePath \"\"" Dec 03 00:11:25 crc kubenswrapper[4953]: I1203 00:11:25.215782 4953 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-ct49l"] Dec 03 00:11:25 crc kubenswrapper[4953]: I1203 00:11:25.218601 4953 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-ct49l"] Dec 03 00:11:25 crc kubenswrapper[4953]: I1203 00:11:25.356478 4953 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3e5a99ac-9abf-4291-b773-fda0c4364959" path="/var/lib/kubelet/pods/3e5a99ac-9abf-4291-b773-fda0c4364959/volumes" Dec 03 00:11:26 crc kubenswrapper[4953]: I1203 00:11:26.396664 4953 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Dec 03 00:11:26 crc kubenswrapper[4953]: E1203 00:11:26.397030 4953 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3e5a99ac-9abf-4291-b773-fda0c4364959" containerName="extract-utilities" Dec 03 00:11:26 crc kubenswrapper[4953]: I1203 00:11:26.397053 4953 state_mem.go:107] "Deleted CPUSet assignment" podUID="3e5a99ac-9abf-4291-b773-fda0c4364959" containerName="extract-utilities" Dec 03 00:11:26 crc kubenswrapper[4953]: E1203 00:11:26.397076 4953 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="be2fa7e6-200e-44f0-9b40-57ed2d30db95" containerName="extract-utilities" Dec 03 00:11:26 crc kubenswrapper[4953]: I1203 00:11:26.397092 4953 state_mem.go:107] "Deleted CPUSet assignment" podUID="be2fa7e6-200e-44f0-9b40-57ed2d30db95" containerName="extract-utilities" Dec 03 00:11:26 crc kubenswrapper[4953]: E1203 00:11:26.397116 4953 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3e5a99ac-9abf-4291-b773-fda0c4364959" containerName="registry-server" Dec 03 00:11:26 crc kubenswrapper[4953]: I1203 00:11:26.397128 4953 state_mem.go:107] "Deleted CPUSet assignment" podUID="3e5a99ac-9abf-4291-b773-fda0c4364959" containerName="registry-server" Dec 03 00:11:26 crc kubenswrapper[4953]: E1203 00:11:26.397146 4953 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="496add53-6511-4560-82c5-3efeb65935f0" containerName="extract-utilities" Dec 03 00:11:26 crc kubenswrapper[4953]: I1203 00:11:26.397159 4953 state_mem.go:107] "Deleted CPUSet assignment" podUID="496add53-6511-4560-82c5-3efeb65935f0" containerName="extract-utilities" Dec 03 00:11:26 crc kubenswrapper[4953]: E1203 00:11:26.397175 4953 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="496add53-6511-4560-82c5-3efeb65935f0" containerName="registry-server" Dec 03 00:11:26 crc kubenswrapper[4953]: I1203 00:11:26.397186 4953 state_mem.go:107] "Deleted CPUSet assignment" podUID="496add53-6511-4560-82c5-3efeb65935f0" containerName="registry-server" Dec 03 00:11:26 crc kubenswrapper[4953]: E1203 00:11:26.397206 4953 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="be2fa7e6-200e-44f0-9b40-57ed2d30db95" containerName="registry-server" Dec 03 00:11:26 crc kubenswrapper[4953]: I1203 00:11:26.397217 4953 state_mem.go:107] "Deleted CPUSet assignment" podUID="be2fa7e6-200e-44f0-9b40-57ed2d30db95" containerName="registry-server" Dec 03 00:11:26 crc kubenswrapper[4953]: E1203 00:11:26.397231 4953 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f17dcaea-cb22-4359-9480-e1d129cd309e" containerName="extract-content" Dec 03 00:11:26 crc kubenswrapper[4953]: I1203 00:11:26.397243 4953 state_mem.go:107] "Deleted CPUSet assignment" podUID="f17dcaea-cb22-4359-9480-e1d129cd309e" containerName="extract-content" Dec 03 00:11:26 crc kubenswrapper[4953]: E1203 00:11:26.397262 4953 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f17dcaea-cb22-4359-9480-e1d129cd309e" containerName="extract-utilities" Dec 03 00:11:26 crc kubenswrapper[4953]: I1203 00:11:26.397276 4953 state_mem.go:107] "Deleted CPUSet assignment" podUID="f17dcaea-cb22-4359-9480-e1d129cd309e" containerName="extract-utilities" Dec 03 00:11:26 crc kubenswrapper[4953]: E1203 00:11:26.397291 4953 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f17dcaea-cb22-4359-9480-e1d129cd309e" containerName="registry-server" Dec 03 00:11:26 crc kubenswrapper[4953]: I1203 00:11:26.397303 4953 state_mem.go:107] "Deleted CPUSet assignment" podUID="f17dcaea-cb22-4359-9480-e1d129cd309e" containerName="registry-server" Dec 03 00:11:26 crc kubenswrapper[4953]: E1203 00:11:26.397320 4953 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="be2fa7e6-200e-44f0-9b40-57ed2d30db95" containerName="extract-content" Dec 03 00:11:26 crc kubenswrapper[4953]: I1203 00:11:26.397331 4953 state_mem.go:107] "Deleted CPUSet assignment" podUID="be2fa7e6-200e-44f0-9b40-57ed2d30db95" containerName="extract-content" Dec 03 00:11:26 crc kubenswrapper[4953]: E1203 00:11:26.397347 4953 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3e5a99ac-9abf-4291-b773-fda0c4364959" containerName="extract-content" Dec 03 00:11:26 crc kubenswrapper[4953]: I1203 00:11:26.397359 4953 state_mem.go:107] "Deleted CPUSet assignment" podUID="3e5a99ac-9abf-4291-b773-fda0c4364959" containerName="extract-content" Dec 03 00:11:26 crc kubenswrapper[4953]: E1203 00:11:26.397377 4953 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="622700d5-0094-4bd6-b12a-8fb43581b5ab" containerName="pruner" Dec 03 00:11:26 crc kubenswrapper[4953]: I1203 00:11:26.397389 4953 state_mem.go:107] "Deleted CPUSet assignment" podUID="622700d5-0094-4bd6-b12a-8fb43581b5ab" containerName="pruner" Dec 03 00:11:26 crc kubenswrapper[4953]: E1203 00:11:26.397408 4953 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="496add53-6511-4560-82c5-3efeb65935f0" containerName="extract-content" Dec 03 00:11:26 crc kubenswrapper[4953]: I1203 00:11:26.397421 4953 state_mem.go:107] "Deleted CPUSet assignment" podUID="496add53-6511-4560-82c5-3efeb65935f0" containerName="extract-content" Dec 03 00:11:26 crc kubenswrapper[4953]: I1203 00:11:26.397593 4953 memory_manager.go:354] "RemoveStaleState removing state" podUID="f17dcaea-cb22-4359-9480-e1d129cd309e" containerName="registry-server" Dec 03 00:11:26 crc kubenswrapper[4953]: I1203 00:11:26.397612 4953 memory_manager.go:354] "RemoveStaleState removing state" podUID="3e5a99ac-9abf-4291-b773-fda0c4364959" containerName="registry-server" Dec 03 00:11:26 crc kubenswrapper[4953]: I1203 00:11:26.397638 4953 memory_manager.go:354] "RemoveStaleState removing state" podUID="be2fa7e6-200e-44f0-9b40-57ed2d30db95" containerName="registry-server" Dec 03 00:11:26 crc kubenswrapper[4953]: I1203 00:11:26.397658 4953 memory_manager.go:354] "RemoveStaleState removing state" podUID="496add53-6511-4560-82c5-3efeb65935f0" containerName="registry-server" Dec 03 00:11:26 crc kubenswrapper[4953]: I1203 00:11:26.397679 4953 memory_manager.go:354] "RemoveStaleState removing state" podUID="622700d5-0094-4bd6-b12a-8fb43581b5ab" containerName="pruner" Dec 03 00:11:26 crc kubenswrapper[4953]: I1203 00:11:26.399309 4953 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Dec 03 00:11:26 crc kubenswrapper[4953]: I1203 00:11:26.399564 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 03 00:11:26 crc kubenswrapper[4953]: I1203 00:11:26.400117 4953 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" containerID="cri-o://034e0bac6892d361700b8f33d6427b8a55e2bd3f89b14ea77a374c6be94d25e2" gracePeriod=15 Dec 03 00:11:26 crc kubenswrapper[4953]: I1203 00:11:26.400141 4953 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" containerID="cri-o://29e4b9d1b70a93fb133e0759ad148ae133766a66efeae3ff06a21dfae0e805ff" gracePeriod=15 Dec 03 00:11:26 crc kubenswrapper[4953]: I1203 00:11:26.400388 4953 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" containerID="cri-o://3c143d4232fb8bf7ef5999a664e431a00ab0db030f4d31a28d0bb21e00e38581" gracePeriod=15 Dec 03 00:11:26 crc kubenswrapper[4953]: I1203 00:11:26.400448 4953 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" containerID="cri-o://dec43536772aa7c3f5a662fa7f3bd73b1cae2791ad454f85792287828d0ffeed" gracePeriod=15 Dec 03 00:11:26 crc kubenswrapper[4953]: I1203 00:11:26.400041 4953 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" containerID="cri-o://cae5e2994fcc24dc9b2ea73e79572615b3311edc42642b7f278fb84123989b23" gracePeriod=15 Dec 03 00:11:26 crc kubenswrapper[4953]: I1203 00:11:26.401203 4953 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Dec 03 00:11:26 crc kubenswrapper[4953]: E1203 00:11:26.401503 4953 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Dec 03 00:11:26 crc kubenswrapper[4953]: I1203 00:11:26.401527 4953 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Dec 03 00:11:26 crc kubenswrapper[4953]: E1203 00:11:26.401544 4953 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 03 00:11:26 crc kubenswrapper[4953]: I1203 00:11:26.401559 4953 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 03 00:11:26 crc kubenswrapper[4953]: E1203 00:11:26.401580 4953 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Dec 03 00:11:26 crc kubenswrapper[4953]: I1203 00:11:26.401598 4953 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Dec 03 00:11:26 crc kubenswrapper[4953]: E1203 00:11:26.401624 4953 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Dec 03 00:11:26 crc kubenswrapper[4953]: I1203 00:11:26.401638 4953 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Dec 03 00:11:26 crc kubenswrapper[4953]: E1203 00:11:26.401658 4953 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Dec 03 00:11:26 crc kubenswrapper[4953]: I1203 00:11:26.401673 4953 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Dec 03 00:11:26 crc kubenswrapper[4953]: E1203 00:11:26.401697 4953 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Dec 03 00:11:26 crc kubenswrapper[4953]: I1203 00:11:26.401711 4953 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Dec 03 00:11:26 crc kubenswrapper[4953]: I1203 00:11:26.401938 4953 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 03 00:11:26 crc kubenswrapper[4953]: I1203 00:11:26.401968 4953 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Dec 03 00:11:26 crc kubenswrapper[4953]: I1203 00:11:26.402023 4953 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 03 00:11:26 crc kubenswrapper[4953]: I1203 00:11:26.402045 4953 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Dec 03 00:11:26 crc kubenswrapper[4953]: I1203 00:11:26.402061 4953 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Dec 03 00:11:26 crc kubenswrapper[4953]: I1203 00:11:26.402085 4953 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Dec 03 00:11:26 crc kubenswrapper[4953]: E1203 00:11:26.402300 4953 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 03 00:11:26 crc kubenswrapper[4953]: I1203 00:11:26.402320 4953 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 03 00:11:26 crc kubenswrapper[4953]: I1203 00:11:26.586612 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 03 00:11:26 crc kubenswrapper[4953]: I1203 00:11:26.587126 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 03 00:11:26 crc kubenswrapper[4953]: I1203 00:11:26.587158 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 00:11:26 crc kubenswrapper[4953]: I1203 00:11:26.587188 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 03 00:11:26 crc kubenswrapper[4953]: I1203 00:11:26.587212 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 03 00:11:26 crc kubenswrapper[4953]: I1203 00:11:26.587244 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 00:11:26 crc kubenswrapper[4953]: I1203 00:11:26.587272 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 03 00:11:26 crc kubenswrapper[4953]: I1203 00:11:26.587297 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 00:11:26 crc kubenswrapper[4953]: I1203 00:11:26.689309 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 03 00:11:26 crc kubenswrapper[4953]: I1203 00:11:26.689419 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 03 00:11:26 crc kubenswrapper[4953]: I1203 00:11:26.689456 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 00:11:26 crc kubenswrapper[4953]: I1203 00:11:26.689492 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 03 00:11:26 crc kubenswrapper[4953]: I1203 00:11:26.689525 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 03 00:11:26 crc kubenswrapper[4953]: I1203 00:11:26.689520 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 03 00:11:26 crc kubenswrapper[4953]: I1203 00:11:26.689559 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 03 00:11:26 crc kubenswrapper[4953]: I1203 00:11:26.689520 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 03 00:11:26 crc kubenswrapper[4953]: I1203 00:11:26.689560 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 00:11:26 crc kubenswrapper[4953]: I1203 00:11:26.689614 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 03 00:11:26 crc kubenswrapper[4953]: I1203 00:11:26.689614 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 00:11:26 crc kubenswrapper[4953]: I1203 00:11:26.689685 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 00:11:26 crc kubenswrapper[4953]: I1203 00:11:26.689753 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 03 00:11:26 crc kubenswrapper[4953]: I1203 00:11:26.689797 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 00:11:26 crc kubenswrapper[4953]: I1203 00:11:26.689870 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 00:11:26 crc kubenswrapper[4953]: I1203 00:11:26.689866 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 03 00:11:26 crc kubenswrapper[4953]: I1203 00:11:26.894705 4953 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Dec 03 00:11:26 crc kubenswrapper[4953]: I1203 00:11:26.896462 4953 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Dec 03 00:11:26 crc kubenswrapper[4953]: I1203 00:11:26.897456 4953 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="034e0bac6892d361700b8f33d6427b8a55e2bd3f89b14ea77a374c6be94d25e2" exitCode=0 Dec 03 00:11:26 crc kubenswrapper[4953]: I1203 00:11:26.897572 4953 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="3c143d4232fb8bf7ef5999a664e431a00ab0db030f4d31a28d0bb21e00e38581" exitCode=0 Dec 03 00:11:26 crc kubenswrapper[4953]: I1203 00:11:26.897573 4953 scope.go:117] "RemoveContainer" containerID="761973393a93c59c23e69f41b8ac0d16abc35abd61ecbc6d51a42d94cd664233" Dec 03 00:11:26 crc kubenswrapper[4953]: I1203 00:11:26.897665 4953 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="29e4b9d1b70a93fb133e0759ad148ae133766a66efeae3ff06a21dfae0e805ff" exitCode=0 Dec 03 00:11:26 crc kubenswrapper[4953]: I1203 00:11:26.897848 4953 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="dec43536772aa7c3f5a662fa7f3bd73b1cae2791ad454f85792287828d0ffeed" exitCode=2 Dec 03 00:11:26 crc kubenswrapper[4953]: I1203 00:11:26.900718 4953 generic.go:334] "Generic (PLEG): container finished" podID="bc833625-0910-40d3-8f2d-801b4b9c40d2" containerID="bebd17b4416c9e8f950db20eecd3806544c7aa8c0f46296470a6535c84598c5a" exitCode=0 Dec 03 00:11:26 crc kubenswrapper[4953]: I1203 00:11:26.900766 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"bc833625-0910-40d3-8f2d-801b4b9c40d2","Type":"ContainerDied","Data":"bebd17b4416c9e8f950db20eecd3806544c7aa8c0f46296470a6535c84598c5a"} Dec 03 00:11:26 crc kubenswrapper[4953]: I1203 00:11:26.902036 4953 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.158:6443: connect: connection refused" Dec 03 00:11:26 crc kubenswrapper[4953]: I1203 00:11:26.902591 4953 status_manager.go:851] "Failed to get status for pod" podUID="bc833625-0910-40d3-8f2d-801b4b9c40d2" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.158:6443: connect: connection refused" Dec 03 00:11:27 crc kubenswrapper[4953]: I1203 00:11:27.302146 4953 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:6443/readyz\": dial tcp 192.168.126.11:6443: connect: connection refused" start-of-body= Dec 03 00:11:27 crc kubenswrapper[4953]: I1203 00:11:27.302219 4953 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="Get \"https://192.168.126.11:6443/readyz\": dial tcp 192.168.126.11:6443: connect: connection refused" Dec 03 00:11:27 crc kubenswrapper[4953]: I1203 00:11:27.911231 4953 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Dec 03 00:11:28 crc kubenswrapper[4953]: I1203 00:11:28.155290 4953 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Dec 03 00:11:28 crc kubenswrapper[4953]: I1203 00:11:28.157352 4953 status_manager.go:851] "Failed to get status for pod" podUID="bc833625-0910-40d3-8f2d-801b4b9c40d2" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.158:6443: connect: connection refused" Dec 03 00:11:28 crc kubenswrapper[4953]: I1203 00:11:28.314176 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/bc833625-0910-40d3-8f2d-801b4b9c40d2-var-lock\") pod \"bc833625-0910-40d3-8f2d-801b4b9c40d2\" (UID: \"bc833625-0910-40d3-8f2d-801b4b9c40d2\") " Dec 03 00:11:28 crc kubenswrapper[4953]: I1203 00:11:28.314357 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/bc833625-0910-40d3-8f2d-801b4b9c40d2-kubelet-dir\") pod \"bc833625-0910-40d3-8f2d-801b4b9c40d2\" (UID: \"bc833625-0910-40d3-8f2d-801b4b9c40d2\") " Dec 03 00:11:28 crc kubenswrapper[4953]: I1203 00:11:28.314372 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/bc833625-0910-40d3-8f2d-801b4b9c40d2-var-lock" (OuterVolumeSpecName: "var-lock") pod "bc833625-0910-40d3-8f2d-801b4b9c40d2" (UID: "bc833625-0910-40d3-8f2d-801b4b9c40d2"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 00:11:28 crc kubenswrapper[4953]: I1203 00:11:28.314391 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/bc833625-0910-40d3-8f2d-801b4b9c40d2-kube-api-access\") pod \"bc833625-0910-40d3-8f2d-801b4b9c40d2\" (UID: \"bc833625-0910-40d3-8f2d-801b4b9c40d2\") " Dec 03 00:11:28 crc kubenswrapper[4953]: I1203 00:11:28.314435 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/bc833625-0910-40d3-8f2d-801b4b9c40d2-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "bc833625-0910-40d3-8f2d-801b4b9c40d2" (UID: "bc833625-0910-40d3-8f2d-801b4b9c40d2"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 00:11:28 crc kubenswrapper[4953]: I1203 00:11:28.314719 4953 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/bc833625-0910-40d3-8f2d-801b4b9c40d2-var-lock\") on node \"crc\" DevicePath \"\"" Dec 03 00:11:28 crc kubenswrapper[4953]: I1203 00:11:28.314734 4953 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/bc833625-0910-40d3-8f2d-801b4b9c40d2-kubelet-dir\") on node \"crc\" DevicePath \"\"" Dec 03 00:11:28 crc kubenswrapper[4953]: I1203 00:11:28.324369 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc833625-0910-40d3-8f2d-801b4b9c40d2-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "bc833625-0910-40d3-8f2d-801b4b9c40d2" (UID: "bc833625-0910-40d3-8f2d-801b4b9c40d2"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:11:28 crc kubenswrapper[4953]: I1203 00:11:28.416620 4953 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/bc833625-0910-40d3-8f2d-801b4b9c40d2-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 03 00:11:28 crc kubenswrapper[4953]: I1203 00:11:28.873261 4953 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Dec 03 00:11:28 crc kubenswrapper[4953]: I1203 00:11:28.874513 4953 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 00:11:28 crc kubenswrapper[4953]: I1203 00:11:28.875434 4953 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.158:6443: connect: connection refused" Dec 03 00:11:28 crc kubenswrapper[4953]: I1203 00:11:28.876232 4953 status_manager.go:851] "Failed to get status for pod" podUID="bc833625-0910-40d3-8f2d-801b4b9c40d2" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.158:6443: connect: connection refused" Dec 03 00:11:28 crc kubenswrapper[4953]: I1203 00:11:28.921011 4953 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Dec 03 00:11:28 crc kubenswrapper[4953]: I1203 00:11:28.922022 4953 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="cae5e2994fcc24dc9b2ea73e79572615b3311edc42642b7f278fb84123989b23" exitCode=0 Dec 03 00:11:28 crc kubenswrapper[4953]: I1203 00:11:28.922129 4953 scope.go:117] "RemoveContainer" containerID="034e0bac6892d361700b8f33d6427b8a55e2bd3f89b14ea77a374c6be94d25e2" Dec 03 00:11:28 crc kubenswrapper[4953]: I1203 00:11:28.922173 4953 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 00:11:28 crc kubenswrapper[4953]: I1203 00:11:28.924372 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"bc833625-0910-40d3-8f2d-801b4b9c40d2","Type":"ContainerDied","Data":"e4ef29be021e5aec33a5212e8eb58d157536c4efd345782f17b4cbd4069a3cb7"} Dec 03 00:11:28 crc kubenswrapper[4953]: I1203 00:11:28.924418 4953 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e4ef29be021e5aec33a5212e8eb58d157536c4efd345782f17b4cbd4069a3cb7" Dec 03 00:11:28 crc kubenswrapper[4953]: I1203 00:11:28.924421 4953 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Dec 03 00:11:28 crc kubenswrapper[4953]: I1203 00:11:28.939872 4953 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.158:6443: connect: connection refused" Dec 03 00:11:28 crc kubenswrapper[4953]: I1203 00:11:28.940536 4953 status_manager.go:851] "Failed to get status for pod" podUID="bc833625-0910-40d3-8f2d-801b4b9c40d2" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.158:6443: connect: connection refused" Dec 03 00:11:28 crc kubenswrapper[4953]: I1203 00:11:28.944177 4953 scope.go:117] "RemoveContainer" containerID="3c143d4232fb8bf7ef5999a664e431a00ab0db030f4d31a28d0bb21e00e38581" Dec 03 00:11:28 crc kubenswrapper[4953]: I1203 00:11:28.963179 4953 scope.go:117] "RemoveContainer" containerID="29e4b9d1b70a93fb133e0759ad148ae133766a66efeae3ff06a21dfae0e805ff" Dec 03 00:11:28 crc kubenswrapper[4953]: I1203 00:11:28.986952 4953 scope.go:117] "RemoveContainer" containerID="dec43536772aa7c3f5a662fa7f3bd73b1cae2791ad454f85792287828d0ffeed" Dec 03 00:11:28 crc kubenswrapper[4953]: E1203 00:11:28.987667 4953 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.158:6443: connect: connection refused" Dec 03 00:11:28 crc kubenswrapper[4953]: E1203 00:11:28.987897 4953 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.158:6443: connect: connection refused" Dec 03 00:11:28 crc kubenswrapper[4953]: E1203 00:11:28.988226 4953 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.158:6443: connect: connection refused" Dec 03 00:11:28 crc kubenswrapper[4953]: E1203 00:11:28.988576 4953 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.158:6443: connect: connection refused" Dec 03 00:11:28 crc kubenswrapper[4953]: E1203 00:11:28.988894 4953 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.158:6443: connect: connection refused" Dec 03 00:11:28 crc kubenswrapper[4953]: I1203 00:11:28.988958 4953 controller.go:115] "failed to update lease using latest lease, fallback to ensure lease" err="failed 5 attempts to update lease" Dec 03 00:11:28 crc kubenswrapper[4953]: E1203 00:11:28.989431 4953 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.158:6443: connect: connection refused" interval="200ms" Dec 03 00:11:29 crc kubenswrapper[4953]: I1203 00:11:29.006229 4953 scope.go:117] "RemoveContainer" containerID="cae5e2994fcc24dc9b2ea73e79572615b3311edc42642b7f278fb84123989b23" Dec 03 00:11:29 crc kubenswrapper[4953]: I1203 00:11:29.024140 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Dec 03 00:11:29 crc kubenswrapper[4953]: I1203 00:11:29.024231 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir" (OuterVolumeSpecName: "cert-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "cert-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 00:11:29 crc kubenswrapper[4953]: I1203 00:11:29.024428 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Dec 03 00:11:29 crc kubenswrapper[4953]: I1203 00:11:29.024471 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Dec 03 00:11:29 crc kubenswrapper[4953]: I1203 00:11:29.024577 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 00:11:29 crc kubenswrapper[4953]: I1203 00:11:29.024603 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 00:11:29 crc kubenswrapper[4953]: I1203 00:11:29.024889 4953 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") on node \"crc\" DevicePath \"\"" Dec 03 00:11:29 crc kubenswrapper[4953]: I1203 00:11:29.024910 4953 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") on node \"crc\" DevicePath \"\"" Dec 03 00:11:29 crc kubenswrapper[4953]: I1203 00:11:29.024918 4953 reconciler_common.go:293] "Volume detached for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") on node \"crc\" DevicePath \"\"" Dec 03 00:11:29 crc kubenswrapper[4953]: I1203 00:11:29.029209 4953 scope.go:117] "RemoveContainer" containerID="72af6c27296eae2e275d455795c68f0d272869d63642b8353b17d466d0c901cf" Dec 03 00:11:29 crc kubenswrapper[4953]: I1203 00:11:29.054969 4953 scope.go:117] "RemoveContainer" containerID="034e0bac6892d361700b8f33d6427b8a55e2bd3f89b14ea77a374c6be94d25e2" Dec 03 00:11:29 crc kubenswrapper[4953]: E1203 00:11:29.055545 4953 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"034e0bac6892d361700b8f33d6427b8a55e2bd3f89b14ea77a374c6be94d25e2\": container with ID starting with 034e0bac6892d361700b8f33d6427b8a55e2bd3f89b14ea77a374c6be94d25e2 not found: ID does not exist" containerID="034e0bac6892d361700b8f33d6427b8a55e2bd3f89b14ea77a374c6be94d25e2" Dec 03 00:11:29 crc kubenswrapper[4953]: I1203 00:11:29.055615 4953 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"034e0bac6892d361700b8f33d6427b8a55e2bd3f89b14ea77a374c6be94d25e2"} err="failed to get container status \"034e0bac6892d361700b8f33d6427b8a55e2bd3f89b14ea77a374c6be94d25e2\": rpc error: code = NotFound desc = could not find container \"034e0bac6892d361700b8f33d6427b8a55e2bd3f89b14ea77a374c6be94d25e2\": container with ID starting with 034e0bac6892d361700b8f33d6427b8a55e2bd3f89b14ea77a374c6be94d25e2 not found: ID does not exist" Dec 03 00:11:29 crc kubenswrapper[4953]: I1203 00:11:29.055650 4953 scope.go:117] "RemoveContainer" containerID="3c143d4232fb8bf7ef5999a664e431a00ab0db030f4d31a28d0bb21e00e38581" Dec 03 00:11:29 crc kubenswrapper[4953]: E1203 00:11:29.056266 4953 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3c143d4232fb8bf7ef5999a664e431a00ab0db030f4d31a28d0bb21e00e38581\": container with ID starting with 3c143d4232fb8bf7ef5999a664e431a00ab0db030f4d31a28d0bb21e00e38581 not found: ID does not exist" containerID="3c143d4232fb8bf7ef5999a664e431a00ab0db030f4d31a28d0bb21e00e38581" Dec 03 00:11:29 crc kubenswrapper[4953]: I1203 00:11:29.056306 4953 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3c143d4232fb8bf7ef5999a664e431a00ab0db030f4d31a28d0bb21e00e38581"} err="failed to get container status \"3c143d4232fb8bf7ef5999a664e431a00ab0db030f4d31a28d0bb21e00e38581\": rpc error: code = NotFound desc = could not find container \"3c143d4232fb8bf7ef5999a664e431a00ab0db030f4d31a28d0bb21e00e38581\": container with ID starting with 3c143d4232fb8bf7ef5999a664e431a00ab0db030f4d31a28d0bb21e00e38581 not found: ID does not exist" Dec 03 00:11:29 crc kubenswrapper[4953]: I1203 00:11:29.056336 4953 scope.go:117] "RemoveContainer" containerID="29e4b9d1b70a93fb133e0759ad148ae133766a66efeae3ff06a21dfae0e805ff" Dec 03 00:11:29 crc kubenswrapper[4953]: E1203 00:11:29.056746 4953 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"29e4b9d1b70a93fb133e0759ad148ae133766a66efeae3ff06a21dfae0e805ff\": container with ID starting with 29e4b9d1b70a93fb133e0759ad148ae133766a66efeae3ff06a21dfae0e805ff not found: ID does not exist" containerID="29e4b9d1b70a93fb133e0759ad148ae133766a66efeae3ff06a21dfae0e805ff" Dec 03 00:11:29 crc kubenswrapper[4953]: I1203 00:11:29.056778 4953 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"29e4b9d1b70a93fb133e0759ad148ae133766a66efeae3ff06a21dfae0e805ff"} err="failed to get container status \"29e4b9d1b70a93fb133e0759ad148ae133766a66efeae3ff06a21dfae0e805ff\": rpc error: code = NotFound desc = could not find container \"29e4b9d1b70a93fb133e0759ad148ae133766a66efeae3ff06a21dfae0e805ff\": container with ID starting with 29e4b9d1b70a93fb133e0759ad148ae133766a66efeae3ff06a21dfae0e805ff not found: ID does not exist" Dec 03 00:11:29 crc kubenswrapper[4953]: I1203 00:11:29.056791 4953 scope.go:117] "RemoveContainer" containerID="dec43536772aa7c3f5a662fa7f3bd73b1cae2791ad454f85792287828d0ffeed" Dec 03 00:11:29 crc kubenswrapper[4953]: E1203 00:11:29.057105 4953 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dec43536772aa7c3f5a662fa7f3bd73b1cae2791ad454f85792287828d0ffeed\": container with ID starting with dec43536772aa7c3f5a662fa7f3bd73b1cae2791ad454f85792287828d0ffeed not found: ID does not exist" containerID="dec43536772aa7c3f5a662fa7f3bd73b1cae2791ad454f85792287828d0ffeed" Dec 03 00:11:29 crc kubenswrapper[4953]: I1203 00:11:29.057128 4953 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dec43536772aa7c3f5a662fa7f3bd73b1cae2791ad454f85792287828d0ffeed"} err="failed to get container status \"dec43536772aa7c3f5a662fa7f3bd73b1cae2791ad454f85792287828d0ffeed\": rpc error: code = NotFound desc = could not find container \"dec43536772aa7c3f5a662fa7f3bd73b1cae2791ad454f85792287828d0ffeed\": container with ID starting with dec43536772aa7c3f5a662fa7f3bd73b1cae2791ad454f85792287828d0ffeed not found: ID does not exist" Dec 03 00:11:29 crc kubenswrapper[4953]: I1203 00:11:29.057140 4953 scope.go:117] "RemoveContainer" containerID="cae5e2994fcc24dc9b2ea73e79572615b3311edc42642b7f278fb84123989b23" Dec 03 00:11:29 crc kubenswrapper[4953]: E1203 00:11:29.057416 4953 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cae5e2994fcc24dc9b2ea73e79572615b3311edc42642b7f278fb84123989b23\": container with ID starting with cae5e2994fcc24dc9b2ea73e79572615b3311edc42642b7f278fb84123989b23 not found: ID does not exist" containerID="cae5e2994fcc24dc9b2ea73e79572615b3311edc42642b7f278fb84123989b23" Dec 03 00:11:29 crc kubenswrapper[4953]: I1203 00:11:29.057437 4953 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cae5e2994fcc24dc9b2ea73e79572615b3311edc42642b7f278fb84123989b23"} err="failed to get container status \"cae5e2994fcc24dc9b2ea73e79572615b3311edc42642b7f278fb84123989b23\": rpc error: code = NotFound desc = could not find container \"cae5e2994fcc24dc9b2ea73e79572615b3311edc42642b7f278fb84123989b23\": container with ID starting with cae5e2994fcc24dc9b2ea73e79572615b3311edc42642b7f278fb84123989b23 not found: ID does not exist" Dec 03 00:11:29 crc kubenswrapper[4953]: I1203 00:11:29.057451 4953 scope.go:117] "RemoveContainer" containerID="72af6c27296eae2e275d455795c68f0d272869d63642b8353b17d466d0c901cf" Dec 03 00:11:29 crc kubenswrapper[4953]: E1203 00:11:29.057699 4953 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"72af6c27296eae2e275d455795c68f0d272869d63642b8353b17d466d0c901cf\": container with ID starting with 72af6c27296eae2e275d455795c68f0d272869d63642b8353b17d466d0c901cf not found: ID does not exist" containerID="72af6c27296eae2e275d455795c68f0d272869d63642b8353b17d466d0c901cf" Dec 03 00:11:29 crc kubenswrapper[4953]: I1203 00:11:29.057734 4953 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"72af6c27296eae2e275d455795c68f0d272869d63642b8353b17d466d0c901cf"} err="failed to get container status \"72af6c27296eae2e275d455795c68f0d272869d63642b8353b17d466d0c901cf\": rpc error: code = NotFound desc = could not find container \"72af6c27296eae2e275d455795c68f0d272869d63642b8353b17d466d0c901cf\": container with ID starting with 72af6c27296eae2e275d455795c68f0d272869d63642b8353b17d466d0c901cf not found: ID does not exist" Dec 03 00:11:29 crc kubenswrapper[4953]: E1203 00:11:29.190336 4953 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.158:6443: connect: connection refused" interval="400ms" Dec 03 00:11:29 crc kubenswrapper[4953]: I1203 00:11:29.236657 4953 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.158:6443: connect: connection refused" Dec 03 00:11:29 crc kubenswrapper[4953]: I1203 00:11:29.237244 4953 status_manager.go:851] "Failed to get status for pod" podUID="bc833625-0910-40d3-8f2d-801b4b9c40d2" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.158:6443: connect: connection refused" Dec 03 00:11:29 crc kubenswrapper[4953]: I1203 00:11:29.355755 4953 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f4b27818a5e8e43d0dc095d08835c792" path="/var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/volumes" Dec 03 00:11:29 crc kubenswrapper[4953]: E1203 00:11:29.591851 4953 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.158:6443: connect: connection refused" interval="800ms" Dec 03 00:11:30 crc kubenswrapper[4953]: E1203 00:11:30.394608 4953 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.158:6443: connect: connection refused" interval="1.6s" Dec 03 00:11:31 crc kubenswrapper[4953]: I1203 00:11:31.352683 4953 status_manager.go:851] "Failed to get status for pod" podUID="bc833625-0910-40d3-8f2d-801b4b9c40d2" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.158:6443: connect: connection refused" Dec 03 00:11:31 crc kubenswrapper[4953]: E1203 00:11:31.453847 4953 kubelet.go:1929] "Failed creating a mirror pod for" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods\": dial tcp 38.102.83.158:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 03 00:11:31 crc kubenswrapper[4953]: I1203 00:11:31.454600 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 03 00:11:31 crc kubenswrapper[4953]: E1203 00:11:31.506395 4953 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/events\": dial tcp 38.102.83.158:6443: connect: connection refused" event="&Event{ObjectMeta:{kube-apiserver-startup-monitor-crc.187d8c1e78a1b1e6 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-startup-monitor-crc,UID:f85e55b1a89d02b0cb034b1ea31ed45a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{startup-monitor},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-03 00:11:31.499872742 +0000 UTC m=+280.484153535,LastTimestamp:2025-12-03 00:11:31.499872742 +0000 UTC m=+280.484153535,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 03 00:11:31 crc kubenswrapper[4953]: I1203 00:11:31.957046 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"c09cfbc07723f81e802a591039a981ac41ceca4a53e918b96d26ec5f5a633ba8"} Dec 03 00:11:31 crc kubenswrapper[4953]: I1203 00:11:31.957780 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"7966cf014fcfc0e4e450c409365b6a3bf4730ac5e3f33eabda4888805ea5ad46"} Dec 03 00:11:32 crc kubenswrapper[4953]: E1203 00:11:31.959125 4953 kubelet.go:1929] "Failed creating a mirror pod for" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods\": dial tcp 38.102.83.158:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 03 00:11:32 crc kubenswrapper[4953]: I1203 00:11:31.959225 4953 status_manager.go:851] "Failed to get status for pod" podUID="bc833625-0910-40d3-8f2d-801b4b9c40d2" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.158:6443: connect: connection refused" Dec 03 00:11:32 crc kubenswrapper[4953]: E1203 00:11:31.995829 4953 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.158:6443: connect: connection refused" interval="3.2s" Dec 03 00:11:35 crc kubenswrapper[4953]: E1203 00:11:35.197111 4953 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.158:6443: connect: connection refused" interval="6.4s" Dec 03 00:11:38 crc kubenswrapper[4953]: E1203 00:11:38.415200 4953 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/events\": dial tcp 38.102.83.158:6443: connect: connection refused" event="&Event{ObjectMeta:{kube-apiserver-startup-monitor-crc.187d8c1e78a1b1e6 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-startup-monitor-crc,UID:f85e55b1a89d02b0cb034b1ea31ed45a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{startup-monitor},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-03 00:11:31.499872742 +0000 UTC m=+280.484153535,LastTimestamp:2025-12-03 00:11:31.499872742 +0000 UTC m=+280.484153535,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 03 00:11:39 crc kubenswrapper[4953]: E1203 00:11:39.046362 4953 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:11:39Z\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:11:39Z\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:11:39Z\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:11:39Z\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Patch \"https://api-int.crc.testing:6443/api/v1/nodes/crc/status?timeout=10s\": dial tcp 38.102.83.158:6443: connect: connection refused" Dec 03 00:11:39 crc kubenswrapper[4953]: E1203 00:11:39.046984 4953 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.102.83.158:6443: connect: connection refused" Dec 03 00:11:39 crc kubenswrapper[4953]: E1203 00:11:39.047467 4953 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.102.83.158:6443: connect: connection refused" Dec 03 00:11:39 crc kubenswrapper[4953]: E1203 00:11:39.047966 4953 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.102.83.158:6443: connect: connection refused" Dec 03 00:11:39 crc kubenswrapper[4953]: E1203 00:11:39.048421 4953 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.102.83.158:6443: connect: connection refused" Dec 03 00:11:39 crc kubenswrapper[4953]: E1203 00:11:39.048456 4953 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 03 00:11:40 crc kubenswrapper[4953]: I1203 00:11:40.641306 4953 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/kube-controller-manager namespace/openshift-kube-controller-manager: Readiness probe status=failure output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" start-of-body= Dec 03 00:11:40 crc kubenswrapper[4953]: I1203 00:11:40.641733 4953 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="kube-controller-manager" probeResult="failure" output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" Dec 03 00:11:41 crc kubenswrapper[4953]: I1203 00:11:41.033476 4953 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Dec 03 00:11:41 crc kubenswrapper[4953]: I1203 00:11:41.033553 4953 generic.go:334] "Generic (PLEG): container finished" podID="f614b9022728cf315e60c057852e563e" containerID="a2810ffc1b141fa7da811406053db9d72c708edf5f2bb1dd75e4d52ab35fd24c" exitCode=1 Dec 03 00:11:41 crc kubenswrapper[4953]: I1203 00:11:41.033599 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerDied","Data":"a2810ffc1b141fa7da811406053db9d72c708edf5f2bb1dd75e4d52ab35fd24c"} Dec 03 00:11:41 crc kubenswrapper[4953]: I1203 00:11:41.034380 4953 scope.go:117] "RemoveContainer" containerID="a2810ffc1b141fa7da811406053db9d72c708edf5f2bb1dd75e4d52ab35fd24c" Dec 03 00:11:41 crc kubenswrapper[4953]: I1203 00:11:41.034709 4953 status_manager.go:851] "Failed to get status for pod" podUID="bc833625-0910-40d3-8f2d-801b4b9c40d2" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.158:6443: connect: connection refused" Dec 03 00:11:41 crc kubenswrapper[4953]: I1203 00:11:41.035368 4953 status_manager.go:851] "Failed to get status for pod" podUID="f614b9022728cf315e60c057852e563e" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-controller-manager/pods/kube-controller-manager-crc\": dial tcp 38.102.83.158:6443: connect: connection refused" Dec 03 00:11:41 crc kubenswrapper[4953]: I1203 00:11:41.347693 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 00:11:41 crc kubenswrapper[4953]: I1203 00:11:41.351236 4953 status_manager.go:851] "Failed to get status for pod" podUID="bc833625-0910-40d3-8f2d-801b4b9c40d2" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.158:6443: connect: connection refused" Dec 03 00:11:41 crc kubenswrapper[4953]: I1203 00:11:41.351904 4953 status_manager.go:851] "Failed to get status for pod" podUID="f614b9022728cf315e60c057852e563e" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-controller-manager/pods/kube-controller-manager-crc\": dial tcp 38.102.83.158:6443: connect: connection refused" Dec 03 00:11:41 crc kubenswrapper[4953]: I1203 00:11:41.352775 4953 status_manager.go:851] "Failed to get status for pod" podUID="bc833625-0910-40d3-8f2d-801b4b9c40d2" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.158:6443: connect: connection refused" Dec 03 00:11:41 crc kubenswrapper[4953]: I1203 00:11:41.353391 4953 status_manager.go:851] "Failed to get status for pod" podUID="f614b9022728cf315e60c057852e563e" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-controller-manager/pods/kube-controller-manager-crc\": dial tcp 38.102.83.158:6443: connect: connection refused" Dec 03 00:11:41 crc kubenswrapper[4953]: I1203 00:11:41.379594 4953 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="cf821840-5353-4620-8f73-99b56b371ab0" Dec 03 00:11:41 crc kubenswrapper[4953]: I1203 00:11:41.379661 4953 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="cf821840-5353-4620-8f73-99b56b371ab0" Dec 03 00:11:41 crc kubenswrapper[4953]: E1203 00:11:41.380336 4953 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.158:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 00:11:41 crc kubenswrapper[4953]: I1203 00:11:41.381057 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 00:11:41 crc kubenswrapper[4953]: W1203 00:11:41.406223 4953 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod71bb4a3aecc4ba5b26c4b7318770ce13.slice/crio-9601b23541f8b2ad7619be0138b1e89d11cb458cdce522ab20c1aa3424225e6d WatchSource:0}: Error finding container 9601b23541f8b2ad7619be0138b1e89d11cb458cdce522ab20c1aa3424225e6d: Status 404 returned error can't find the container with id 9601b23541f8b2ad7619be0138b1e89d11cb458cdce522ab20c1aa3424225e6d Dec 03 00:11:41 crc kubenswrapper[4953]: E1203 00:11:41.598819 4953 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.158:6443: connect: connection refused" interval="7s" Dec 03 00:11:41 crc kubenswrapper[4953]: I1203 00:11:41.860098 4953 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 03 00:11:42 crc kubenswrapper[4953]: I1203 00:11:42.047141 4953 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Dec 03 00:11:42 crc kubenswrapper[4953]: I1203 00:11:42.047322 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"e5cb435394ec9b2b9a14ff9b35ed484b5e81869a007ef498112a84f305efe260"} Dec 03 00:11:42 crc kubenswrapper[4953]: I1203 00:11:42.048886 4953 status_manager.go:851] "Failed to get status for pod" podUID="bc833625-0910-40d3-8f2d-801b4b9c40d2" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.158:6443: connect: connection refused" Dec 03 00:11:42 crc kubenswrapper[4953]: I1203 00:11:42.049372 4953 status_manager.go:851] "Failed to get status for pod" podUID="f614b9022728cf315e60c057852e563e" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-controller-manager/pods/kube-controller-manager-crc\": dial tcp 38.102.83.158:6443: connect: connection refused" Dec 03 00:11:42 crc kubenswrapper[4953]: I1203 00:11:42.049740 4953 generic.go:334] "Generic (PLEG): container finished" podID="71bb4a3aecc4ba5b26c4b7318770ce13" containerID="384cfa514f73d63d65944acb0a9c15e846de7fee6ca07709069ac04a5f290800" exitCode=0 Dec 03 00:11:42 crc kubenswrapper[4953]: I1203 00:11:42.049810 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerDied","Data":"384cfa514f73d63d65944acb0a9c15e846de7fee6ca07709069ac04a5f290800"} Dec 03 00:11:42 crc kubenswrapper[4953]: I1203 00:11:42.049862 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"9601b23541f8b2ad7619be0138b1e89d11cb458cdce522ab20c1aa3424225e6d"} Dec 03 00:11:42 crc kubenswrapper[4953]: I1203 00:11:42.050189 4953 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="cf821840-5353-4620-8f73-99b56b371ab0" Dec 03 00:11:42 crc kubenswrapper[4953]: I1203 00:11:42.050205 4953 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="cf821840-5353-4620-8f73-99b56b371ab0" Dec 03 00:11:42 crc kubenswrapper[4953]: I1203 00:11:42.050656 4953 status_manager.go:851] "Failed to get status for pod" podUID="bc833625-0910-40d3-8f2d-801b4b9c40d2" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.158:6443: connect: connection refused" Dec 03 00:11:42 crc kubenswrapper[4953]: E1203 00:11:42.050683 4953 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.158:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 00:11:42 crc kubenswrapper[4953]: I1203 00:11:42.051410 4953 status_manager.go:851] "Failed to get status for pod" podUID="f614b9022728cf315e60c057852e563e" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-controller-manager/pods/kube-controller-manager-crc\": dial tcp 38.102.83.158:6443: connect: connection refused" Dec 03 00:11:43 crc kubenswrapper[4953]: I1203 00:11:43.060252 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"a2f93fe9d8032058c4856c3f6bfcb5cb5073a6bc840c88c328c99788c2c9d0ce"} Dec 03 00:11:43 crc kubenswrapper[4953]: I1203 00:11:43.060794 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"8946adb07f513a06ebdd64b8c0c9da36fa1c69eb9b8a6f1c732d82142683cc3a"} Dec 03 00:11:43 crc kubenswrapper[4953]: I1203 00:11:43.060815 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"7490adce64fce04a280951c8457bf8d50a6792802361768f097802aa57498641"} Dec 03 00:11:43 crc kubenswrapper[4953]: I1203 00:11:43.899589 4953 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-authentication/oauth-openshift-558db77b4-45k44" podUID="ca879198-0a45-4505-b845-60d8b0f5bb5d" containerName="oauth-openshift" containerID="cri-o://8da38403c4377107fa9a62218c1f47bc8b022b0acc0fc08c14bb8b6bffce93f6" gracePeriod=15 Dec 03 00:11:44 crc kubenswrapper[4953]: I1203 00:11:44.071765 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"2ecdf96aa195723d6c1a2b081e5961c1eb2edecf505990274a88c5c97862d42e"} Dec 03 00:11:44 crc kubenswrapper[4953]: I1203 00:11:44.071824 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"f7556d9fc6b4ee9b1fc1eee4cecab306d912d2c53341add668406e0e06dd9ba4"} Dec 03 00:11:44 crc kubenswrapper[4953]: I1203 00:11:44.072149 4953 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="cf821840-5353-4620-8f73-99b56b371ab0" Dec 03 00:11:44 crc kubenswrapper[4953]: I1203 00:11:44.072164 4953 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="cf821840-5353-4620-8f73-99b56b371ab0" Dec 03 00:11:44 crc kubenswrapper[4953]: I1203 00:11:44.072556 4953 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 00:11:44 crc kubenswrapper[4953]: I1203 00:11:44.077129 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-45k44" event={"ID":"ca879198-0a45-4505-b845-60d8b0f5bb5d","Type":"ContainerDied","Data":"8da38403c4377107fa9a62218c1f47bc8b022b0acc0fc08c14bb8b6bffce93f6"} Dec 03 00:11:44 crc kubenswrapper[4953]: I1203 00:11:44.077078 4953 generic.go:334] "Generic (PLEG): container finished" podID="ca879198-0a45-4505-b845-60d8b0f5bb5d" containerID="8da38403c4377107fa9a62218c1f47bc8b022b0acc0fc08c14bb8b6bffce93f6" exitCode=0 Dec 03 00:11:44 crc kubenswrapper[4953]: I1203 00:11:44.307575 4953 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-45k44" Dec 03 00:11:44 crc kubenswrapper[4953]: I1203 00:11:44.414302 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/ca879198-0a45-4505-b845-60d8b0f5bb5d-audit-dir\") pod \"ca879198-0a45-4505-b845-60d8b0f5bb5d\" (UID: \"ca879198-0a45-4505-b845-60d8b0f5bb5d\") " Dec 03 00:11:44 crc kubenswrapper[4953]: I1203 00:11:44.414398 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/ca879198-0a45-4505-b845-60d8b0f5bb5d-audit-policies\") pod \"ca879198-0a45-4505-b845-60d8b0f5bb5d\" (UID: \"ca879198-0a45-4505-b845-60d8b0f5bb5d\") " Dec 03 00:11:44 crc kubenswrapper[4953]: I1203 00:11:44.414436 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ca879198-0a45-4505-b845-60d8b0f5bb5d-v4-0-config-system-trusted-ca-bundle\") pod \"ca879198-0a45-4505-b845-60d8b0f5bb5d\" (UID: \"ca879198-0a45-4505-b845-60d8b0f5bb5d\") " Dec 03 00:11:44 crc kubenswrapper[4953]: I1203 00:11:44.414430 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ca879198-0a45-4505-b845-60d8b0f5bb5d-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "ca879198-0a45-4505-b845-60d8b0f5bb5d" (UID: "ca879198-0a45-4505-b845-60d8b0f5bb5d"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 00:11:44 crc kubenswrapper[4953]: I1203 00:11:44.414464 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/ca879198-0a45-4505-b845-60d8b0f5bb5d-v4-0-config-system-ocp-branding-template\") pod \"ca879198-0a45-4505-b845-60d8b0f5bb5d\" (UID: \"ca879198-0a45-4505-b845-60d8b0f5bb5d\") " Dec 03 00:11:44 crc kubenswrapper[4953]: I1203 00:11:44.414518 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/ca879198-0a45-4505-b845-60d8b0f5bb5d-v4-0-config-user-template-provider-selection\") pod \"ca879198-0a45-4505-b845-60d8b0f5bb5d\" (UID: \"ca879198-0a45-4505-b845-60d8b0f5bb5d\") " Dec 03 00:11:44 crc kubenswrapper[4953]: I1203 00:11:44.414553 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/ca879198-0a45-4505-b845-60d8b0f5bb5d-v4-0-config-user-idp-0-file-data\") pod \"ca879198-0a45-4505-b845-60d8b0f5bb5d\" (UID: \"ca879198-0a45-4505-b845-60d8b0f5bb5d\") " Dec 03 00:11:44 crc kubenswrapper[4953]: I1203 00:11:44.414585 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/ca879198-0a45-4505-b845-60d8b0f5bb5d-v4-0-config-system-cliconfig\") pod \"ca879198-0a45-4505-b845-60d8b0f5bb5d\" (UID: \"ca879198-0a45-4505-b845-60d8b0f5bb5d\") " Dec 03 00:11:44 crc kubenswrapper[4953]: I1203 00:11:44.414611 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/ca879198-0a45-4505-b845-60d8b0f5bb5d-v4-0-config-system-session\") pod \"ca879198-0a45-4505-b845-60d8b0f5bb5d\" (UID: \"ca879198-0a45-4505-b845-60d8b0f5bb5d\") " Dec 03 00:11:44 crc kubenswrapper[4953]: I1203 00:11:44.414636 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/ca879198-0a45-4505-b845-60d8b0f5bb5d-v4-0-config-user-template-login\") pod \"ca879198-0a45-4505-b845-60d8b0f5bb5d\" (UID: \"ca879198-0a45-4505-b845-60d8b0f5bb5d\") " Dec 03 00:11:44 crc kubenswrapper[4953]: I1203 00:11:44.415567 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ca879198-0a45-4505-b845-60d8b0f5bb5d-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "ca879198-0a45-4505-b845-60d8b0f5bb5d" (UID: "ca879198-0a45-4505-b845-60d8b0f5bb5d"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:11:44 crc kubenswrapper[4953]: I1203 00:11:44.415582 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ca879198-0a45-4505-b845-60d8b0f5bb5d-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "ca879198-0a45-4505-b845-60d8b0f5bb5d" (UID: "ca879198-0a45-4505-b845-60d8b0f5bb5d"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:11:44 crc kubenswrapper[4953]: I1203 00:11:44.416060 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ca879198-0a45-4505-b845-60d8b0f5bb5d-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "ca879198-0a45-4505-b845-60d8b0f5bb5d" (UID: "ca879198-0a45-4505-b845-60d8b0f5bb5d"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:11:44 crc kubenswrapper[4953]: I1203 00:11:44.416168 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/ca879198-0a45-4505-b845-60d8b0f5bb5d-v4-0-config-system-router-certs\") pod \"ca879198-0a45-4505-b845-60d8b0f5bb5d\" (UID: \"ca879198-0a45-4505-b845-60d8b0f5bb5d\") " Dec 03 00:11:44 crc kubenswrapper[4953]: I1203 00:11:44.416208 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/ca879198-0a45-4505-b845-60d8b0f5bb5d-v4-0-config-system-service-ca\") pod \"ca879198-0a45-4505-b845-60d8b0f5bb5d\" (UID: \"ca879198-0a45-4505-b845-60d8b0f5bb5d\") " Dec 03 00:11:44 crc kubenswrapper[4953]: I1203 00:11:44.416265 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2c2jt\" (UniqueName: \"kubernetes.io/projected/ca879198-0a45-4505-b845-60d8b0f5bb5d-kube-api-access-2c2jt\") pod \"ca879198-0a45-4505-b845-60d8b0f5bb5d\" (UID: \"ca879198-0a45-4505-b845-60d8b0f5bb5d\") " Dec 03 00:11:44 crc kubenswrapper[4953]: I1203 00:11:44.416346 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/ca879198-0a45-4505-b845-60d8b0f5bb5d-v4-0-config-user-template-error\") pod \"ca879198-0a45-4505-b845-60d8b0f5bb5d\" (UID: \"ca879198-0a45-4505-b845-60d8b0f5bb5d\") " Dec 03 00:11:44 crc kubenswrapper[4953]: I1203 00:11:44.416424 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/ca879198-0a45-4505-b845-60d8b0f5bb5d-v4-0-config-system-serving-cert\") pod \"ca879198-0a45-4505-b845-60d8b0f5bb5d\" (UID: \"ca879198-0a45-4505-b845-60d8b0f5bb5d\") " Dec 03 00:11:44 crc kubenswrapper[4953]: I1203 00:11:44.417262 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ca879198-0a45-4505-b845-60d8b0f5bb5d-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "ca879198-0a45-4505-b845-60d8b0f5bb5d" (UID: "ca879198-0a45-4505-b845-60d8b0f5bb5d"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:11:44 crc kubenswrapper[4953]: I1203 00:11:44.417316 4953 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/ca879198-0a45-4505-b845-60d8b0f5bb5d-audit-dir\") on node \"crc\" DevicePath \"\"" Dec 03 00:11:44 crc kubenswrapper[4953]: I1203 00:11:44.417339 4953 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/ca879198-0a45-4505-b845-60d8b0f5bb5d-audit-policies\") on node \"crc\" DevicePath \"\"" Dec 03 00:11:44 crc kubenswrapper[4953]: I1203 00:11:44.417349 4953 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ca879198-0a45-4505-b845-60d8b0f5bb5d-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 00:11:44 crc kubenswrapper[4953]: I1203 00:11:44.417358 4953 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/ca879198-0a45-4505-b845-60d8b0f5bb5d-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Dec 03 00:11:44 crc kubenswrapper[4953]: I1203 00:11:44.422438 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ca879198-0a45-4505-b845-60d8b0f5bb5d-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "ca879198-0a45-4505-b845-60d8b0f5bb5d" (UID: "ca879198-0a45-4505-b845-60d8b0f5bb5d"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:11:44 crc kubenswrapper[4953]: I1203 00:11:44.428411 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ca879198-0a45-4505-b845-60d8b0f5bb5d-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "ca879198-0a45-4505-b845-60d8b0f5bb5d" (UID: "ca879198-0a45-4505-b845-60d8b0f5bb5d"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:11:44 crc kubenswrapper[4953]: I1203 00:11:44.428461 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ca879198-0a45-4505-b845-60d8b0f5bb5d-kube-api-access-2c2jt" (OuterVolumeSpecName: "kube-api-access-2c2jt") pod "ca879198-0a45-4505-b845-60d8b0f5bb5d" (UID: "ca879198-0a45-4505-b845-60d8b0f5bb5d"). InnerVolumeSpecName "kube-api-access-2c2jt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:11:44 crc kubenswrapper[4953]: I1203 00:11:44.429906 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ca879198-0a45-4505-b845-60d8b0f5bb5d-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "ca879198-0a45-4505-b845-60d8b0f5bb5d" (UID: "ca879198-0a45-4505-b845-60d8b0f5bb5d"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:11:44 crc kubenswrapper[4953]: I1203 00:11:44.432305 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ca879198-0a45-4505-b845-60d8b0f5bb5d-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "ca879198-0a45-4505-b845-60d8b0f5bb5d" (UID: "ca879198-0a45-4505-b845-60d8b0f5bb5d"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:11:44 crc kubenswrapper[4953]: I1203 00:11:44.432581 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ca879198-0a45-4505-b845-60d8b0f5bb5d-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "ca879198-0a45-4505-b845-60d8b0f5bb5d" (UID: "ca879198-0a45-4505-b845-60d8b0f5bb5d"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:11:44 crc kubenswrapper[4953]: I1203 00:11:44.432740 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ca879198-0a45-4505-b845-60d8b0f5bb5d-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "ca879198-0a45-4505-b845-60d8b0f5bb5d" (UID: "ca879198-0a45-4505-b845-60d8b0f5bb5d"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:11:44 crc kubenswrapper[4953]: I1203 00:11:44.433516 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ca879198-0a45-4505-b845-60d8b0f5bb5d-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "ca879198-0a45-4505-b845-60d8b0f5bb5d" (UID: "ca879198-0a45-4505-b845-60d8b0f5bb5d"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:11:44 crc kubenswrapper[4953]: I1203 00:11:44.434679 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ca879198-0a45-4505-b845-60d8b0f5bb5d-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "ca879198-0a45-4505-b845-60d8b0f5bb5d" (UID: "ca879198-0a45-4505-b845-60d8b0f5bb5d"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:11:44 crc kubenswrapper[4953]: I1203 00:11:44.519338 4953 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/ca879198-0a45-4505-b845-60d8b0f5bb5d-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 00:11:44 crc kubenswrapper[4953]: I1203 00:11:44.519418 4953 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/ca879198-0a45-4505-b845-60d8b0f5bb5d-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Dec 03 00:11:44 crc kubenswrapper[4953]: I1203 00:11:44.519436 4953 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/ca879198-0a45-4505-b845-60d8b0f5bb5d-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Dec 03 00:11:44 crc kubenswrapper[4953]: I1203 00:11:44.519456 4953 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/ca879198-0a45-4505-b845-60d8b0f5bb5d-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Dec 03 00:11:44 crc kubenswrapper[4953]: I1203 00:11:44.519471 4953 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/ca879198-0a45-4505-b845-60d8b0f5bb5d-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Dec 03 00:11:44 crc kubenswrapper[4953]: I1203 00:11:44.519486 4953 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/ca879198-0a45-4505-b845-60d8b0f5bb5d-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Dec 03 00:11:44 crc kubenswrapper[4953]: I1203 00:11:44.519503 4953 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/ca879198-0a45-4505-b845-60d8b0f5bb5d-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Dec 03 00:11:44 crc kubenswrapper[4953]: I1203 00:11:44.519519 4953 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/ca879198-0a45-4505-b845-60d8b0f5bb5d-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Dec 03 00:11:44 crc kubenswrapper[4953]: I1203 00:11:44.519533 4953 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2c2jt\" (UniqueName: \"kubernetes.io/projected/ca879198-0a45-4505-b845-60d8b0f5bb5d-kube-api-access-2c2jt\") on node \"crc\" DevicePath \"\"" Dec 03 00:11:44 crc kubenswrapper[4953]: I1203 00:11:44.519547 4953 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/ca879198-0a45-4505-b845-60d8b0f5bb5d-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Dec 03 00:11:45 crc kubenswrapper[4953]: I1203 00:11:45.100345 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-45k44" event={"ID":"ca879198-0a45-4505-b845-60d8b0f5bb5d","Type":"ContainerDied","Data":"7a2b12fa65694df3b4f7ebafb191325d197976df99a8b1dcf8e7a3a0a15276ed"} Dec 03 00:11:45 crc kubenswrapper[4953]: I1203 00:11:45.100411 4953 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-45k44" Dec 03 00:11:45 crc kubenswrapper[4953]: I1203 00:11:45.100437 4953 scope.go:117] "RemoveContainer" containerID="8da38403c4377107fa9a62218c1f47bc8b022b0acc0fc08c14bb8b6bffce93f6" Dec 03 00:11:46 crc kubenswrapper[4953]: I1203 00:11:46.381426 4953 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 00:11:46 crc kubenswrapper[4953]: I1203 00:11:46.382164 4953 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 00:11:46 crc kubenswrapper[4953]: I1203 00:11:46.393523 4953 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 00:11:49 crc kubenswrapper[4953]: I1203 00:11:49.087657 4953 kubelet.go:1914] "Deleted mirror pod because it is outdated" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 00:11:49 crc kubenswrapper[4953]: I1203 00:11:49.135665 4953 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="cf821840-5353-4620-8f73-99b56b371ab0" Dec 03 00:11:49 crc kubenswrapper[4953]: I1203 00:11:49.135707 4953 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="cf821840-5353-4620-8f73-99b56b371ab0" Dec 03 00:11:49 crc kubenswrapper[4953]: I1203 00:11:49.140536 4953 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 00:11:50 crc kubenswrapper[4953]: I1203 00:11:50.142112 4953 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="cf821840-5353-4620-8f73-99b56b371ab0" Dec 03 00:11:50 crc kubenswrapper[4953]: I1203 00:11:50.142171 4953 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="cf821840-5353-4620-8f73-99b56b371ab0" Dec 03 00:11:50 crc kubenswrapper[4953]: I1203 00:11:50.641136 4953 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 03 00:11:51 crc kubenswrapper[4953]: I1203 00:11:51.372601 4953 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="09a0b145-95f7-48b9-8f7c-7474ee2d67b1" Dec 03 00:11:51 crc kubenswrapper[4953]: I1203 00:11:51.860462 4953 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 03 00:11:51 crc kubenswrapper[4953]: I1203 00:11:51.869773 4953 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 03 00:11:52 crc kubenswrapper[4953]: I1203 00:11:52.162926 4953 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 03 00:11:58 crc kubenswrapper[4953]: I1203 00:11:58.866919 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Dec 03 00:11:58 crc kubenswrapper[4953]: I1203 00:11:58.881203 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Dec 03 00:11:59 crc kubenswrapper[4953]: I1203 00:11:59.113569 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Dec 03 00:11:59 crc kubenswrapper[4953]: I1203 00:11:59.125070 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Dec 03 00:11:59 crc kubenswrapper[4953]: I1203 00:11:59.130678 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Dec 03 00:11:59 crc kubenswrapper[4953]: I1203 00:11:59.285885 4953 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Dec 03 00:11:59 crc kubenswrapper[4953]: I1203 00:11:59.506295 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Dec 03 00:11:59 crc kubenswrapper[4953]: I1203 00:11:59.512006 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Dec 03 00:11:59 crc kubenswrapper[4953]: I1203 00:11:59.653780 4953 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Dec 03 00:11:59 crc kubenswrapper[4953]: I1203 00:11:59.721366 4953 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Dec 03 00:11:59 crc kubenswrapper[4953]: I1203 00:11:59.781905 4953 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Dec 03 00:12:00 crc kubenswrapper[4953]: I1203 00:12:00.145333 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Dec 03 00:12:00 crc kubenswrapper[4953]: I1203 00:12:00.324417 4953 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Dec 03 00:12:00 crc kubenswrapper[4953]: I1203 00:12:00.836030 4953 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Dec 03 00:12:01 crc kubenswrapper[4953]: I1203 00:12:01.099159 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Dec 03 00:12:01 crc kubenswrapper[4953]: I1203 00:12:01.238399 4953 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Dec 03 00:12:01 crc kubenswrapper[4953]: I1203 00:12:01.240403 4953 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Dec 03 00:12:01 crc kubenswrapper[4953]: I1203 00:12:01.244283 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Dec 03 00:12:01 crc kubenswrapper[4953]: I1203 00:12:01.362772 4953 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Dec 03 00:12:01 crc kubenswrapper[4953]: I1203 00:12:01.420337 4953 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Dec 03 00:12:01 crc kubenswrapper[4953]: I1203 00:12:01.486408 4953 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Dec 03 00:12:01 crc kubenswrapper[4953]: I1203 00:12:01.550495 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Dec 03 00:12:01 crc kubenswrapper[4953]: I1203 00:12:01.566631 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Dec 03 00:12:01 crc kubenswrapper[4953]: I1203 00:12:01.567016 4953 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Dec 03 00:12:01 crc kubenswrapper[4953]: I1203 00:12:01.576186 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Dec 03 00:12:01 crc kubenswrapper[4953]: I1203 00:12:01.752752 4953 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Dec 03 00:12:01 crc kubenswrapper[4953]: I1203 00:12:01.852071 4953 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Dec 03 00:12:01 crc kubenswrapper[4953]: I1203 00:12:01.887651 4953 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Dec 03 00:12:02 crc kubenswrapper[4953]: I1203 00:12:02.066682 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Dec 03 00:12:02 crc kubenswrapper[4953]: I1203 00:12:02.163882 4953 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Dec 03 00:12:02 crc kubenswrapper[4953]: I1203 00:12:02.170062 4953 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Dec 03 00:12:02 crc kubenswrapper[4953]: I1203 00:12:02.204019 4953 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Dec 03 00:12:02 crc kubenswrapper[4953]: I1203 00:12:02.358254 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Dec 03 00:12:02 crc kubenswrapper[4953]: I1203 00:12:02.366845 4953 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Dec 03 00:12:02 crc kubenswrapper[4953]: I1203 00:12:02.381269 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Dec 03 00:12:02 crc kubenswrapper[4953]: I1203 00:12:02.394672 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Dec 03 00:12:02 crc kubenswrapper[4953]: I1203 00:12:02.634241 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Dec 03 00:12:02 crc kubenswrapper[4953]: I1203 00:12:02.663240 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Dec 03 00:12:02 crc kubenswrapper[4953]: I1203 00:12:02.744901 4953 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Dec 03 00:12:02 crc kubenswrapper[4953]: I1203 00:12:02.759094 4953 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Dec 03 00:12:02 crc kubenswrapper[4953]: I1203 00:12:02.823458 4953 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Dec 03 00:12:03 crc kubenswrapper[4953]: I1203 00:12:03.008798 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Dec 03 00:12:03 crc kubenswrapper[4953]: I1203 00:12:03.072532 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Dec 03 00:12:03 crc kubenswrapper[4953]: I1203 00:12:03.138555 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Dec 03 00:12:03 crc kubenswrapper[4953]: I1203 00:12:03.140747 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Dec 03 00:12:03 crc kubenswrapper[4953]: I1203 00:12:03.164183 4953 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Dec 03 00:12:03 crc kubenswrapper[4953]: I1203 00:12:03.180257 4953 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Dec 03 00:12:03 crc kubenswrapper[4953]: I1203 00:12:03.239129 4953 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Dec 03 00:12:03 crc kubenswrapper[4953]: I1203 00:12:03.254853 4953 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Dec 03 00:12:03 crc kubenswrapper[4953]: I1203 00:12:03.322636 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Dec 03 00:12:03 crc kubenswrapper[4953]: I1203 00:12:03.354680 4953 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Dec 03 00:12:03 crc kubenswrapper[4953]: I1203 00:12:03.434638 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Dec 03 00:12:03 crc kubenswrapper[4953]: I1203 00:12:03.447562 4953 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Dec 03 00:12:03 crc kubenswrapper[4953]: I1203 00:12:03.494754 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Dec 03 00:12:03 crc kubenswrapper[4953]: I1203 00:12:03.802341 4953 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Dec 03 00:12:03 crc kubenswrapper[4953]: I1203 00:12:03.805829 4953 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Dec 03 00:12:03 crc kubenswrapper[4953]: I1203 00:12:03.838322 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Dec 03 00:12:04 crc kubenswrapper[4953]: I1203 00:12:04.019095 4953 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Dec 03 00:12:04 crc kubenswrapper[4953]: I1203 00:12:04.031852 4953 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Dec 03 00:12:04 crc kubenswrapper[4953]: I1203 00:12:04.083168 4953 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Dec 03 00:12:04 crc kubenswrapper[4953]: I1203 00:12:04.223853 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Dec 03 00:12:04 crc kubenswrapper[4953]: I1203 00:12:04.270767 4953 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Dec 03 00:12:04 crc kubenswrapper[4953]: I1203 00:12:04.276527 4953 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc","openshift-authentication/oauth-openshift-558db77b4-45k44"] Dec 03 00:12:04 crc kubenswrapper[4953]: I1203 00:12:04.276639 4953 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Dec 03 00:12:04 crc kubenswrapper[4953]: I1203 00:12:04.284292 4953 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 00:12:04 crc kubenswrapper[4953]: I1203 00:12:04.299516 4953 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=15.299497903 podStartE2EDuration="15.299497903s" podCreationTimestamp="2025-12-03 00:11:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 00:12:04.296852366 +0000 UTC m=+313.281133149" watchObservedRunningTime="2025-12-03 00:12:04.299497903 +0000 UTC m=+313.283778686" Dec 03 00:12:04 crc kubenswrapper[4953]: I1203 00:12:04.461386 4953 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Dec 03 00:12:04 crc kubenswrapper[4953]: I1203 00:12:04.468815 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Dec 03 00:12:04 crc kubenswrapper[4953]: I1203 00:12:04.659008 4953 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Dec 03 00:12:05 crc kubenswrapper[4953]: I1203 00:12:05.062730 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Dec 03 00:12:05 crc kubenswrapper[4953]: I1203 00:12:05.100856 4953 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Dec 03 00:12:05 crc kubenswrapper[4953]: I1203 00:12:05.155534 4953 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Dec 03 00:12:05 crc kubenswrapper[4953]: I1203 00:12:05.207341 4953 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Dec 03 00:12:05 crc kubenswrapper[4953]: I1203 00:12:05.277879 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Dec 03 00:12:05 crc kubenswrapper[4953]: I1203 00:12:05.356924 4953 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ca879198-0a45-4505-b845-60d8b0f5bb5d" path="/var/lib/kubelet/pods/ca879198-0a45-4505-b845-60d8b0f5bb5d/volumes" Dec 03 00:12:05 crc kubenswrapper[4953]: I1203 00:12:05.363873 4953 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Dec 03 00:12:05 crc kubenswrapper[4953]: I1203 00:12:05.446073 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Dec 03 00:12:05 crc kubenswrapper[4953]: I1203 00:12:05.453551 4953 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Dec 03 00:12:05 crc kubenswrapper[4953]: I1203 00:12:05.454254 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Dec 03 00:12:05 crc kubenswrapper[4953]: I1203 00:12:05.459106 4953 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Dec 03 00:12:05 crc kubenswrapper[4953]: I1203 00:12:05.512085 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Dec 03 00:12:05 crc kubenswrapper[4953]: I1203 00:12:05.617609 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Dec 03 00:12:05 crc kubenswrapper[4953]: I1203 00:12:05.635476 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Dec 03 00:12:05 crc kubenswrapper[4953]: I1203 00:12:05.661949 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Dec 03 00:12:05 crc kubenswrapper[4953]: I1203 00:12:05.678286 4953 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Dec 03 00:12:05 crc kubenswrapper[4953]: I1203 00:12:05.695539 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Dec 03 00:12:05 crc kubenswrapper[4953]: I1203 00:12:05.720084 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Dec 03 00:12:05 crc kubenswrapper[4953]: I1203 00:12:05.750812 4953 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Dec 03 00:12:05 crc kubenswrapper[4953]: I1203 00:12:05.903273 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Dec 03 00:12:05 crc kubenswrapper[4953]: I1203 00:12:05.939964 4953 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Dec 03 00:12:06 crc kubenswrapper[4953]: I1203 00:12:06.000303 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Dec 03 00:12:06 crc kubenswrapper[4953]: I1203 00:12:06.036538 4953 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Dec 03 00:12:06 crc kubenswrapper[4953]: I1203 00:12:06.084625 4953 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Dec 03 00:12:06 crc kubenswrapper[4953]: I1203 00:12:06.114389 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Dec 03 00:12:06 crc kubenswrapper[4953]: I1203 00:12:06.136009 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Dec 03 00:12:06 crc kubenswrapper[4953]: I1203 00:12:06.146166 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Dec 03 00:12:06 crc kubenswrapper[4953]: I1203 00:12:06.163435 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Dec 03 00:12:06 crc kubenswrapper[4953]: I1203 00:12:06.165779 4953 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Dec 03 00:12:06 crc kubenswrapper[4953]: I1203 00:12:06.229671 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Dec 03 00:12:06 crc kubenswrapper[4953]: I1203 00:12:06.243104 4953 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Dec 03 00:12:06 crc kubenswrapper[4953]: I1203 00:12:06.298599 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Dec 03 00:12:06 crc kubenswrapper[4953]: I1203 00:12:06.355244 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Dec 03 00:12:06 crc kubenswrapper[4953]: I1203 00:12:06.370726 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Dec 03 00:12:06 crc kubenswrapper[4953]: I1203 00:12:06.477687 4953 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Dec 03 00:12:06 crc kubenswrapper[4953]: I1203 00:12:06.527232 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Dec 03 00:12:06 crc kubenswrapper[4953]: I1203 00:12:06.582871 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Dec 03 00:12:06 crc kubenswrapper[4953]: I1203 00:12:06.585090 4953 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Dec 03 00:12:06 crc kubenswrapper[4953]: I1203 00:12:06.589780 4953 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Dec 03 00:12:06 crc kubenswrapper[4953]: I1203 00:12:06.598585 4953 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Dec 03 00:12:06 crc kubenswrapper[4953]: I1203 00:12:06.603176 4953 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Dec 03 00:12:06 crc kubenswrapper[4953]: I1203 00:12:06.722489 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Dec 03 00:12:06 crc kubenswrapper[4953]: I1203 00:12:06.771208 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Dec 03 00:12:06 crc kubenswrapper[4953]: I1203 00:12:06.801721 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Dec 03 00:12:06 crc kubenswrapper[4953]: I1203 00:12:06.862577 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Dec 03 00:12:06 crc kubenswrapper[4953]: I1203 00:12:06.866316 4953 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Dec 03 00:12:06 crc kubenswrapper[4953]: I1203 00:12:06.877450 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Dec 03 00:12:06 crc kubenswrapper[4953]: I1203 00:12:06.927201 4953 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Dec 03 00:12:06 crc kubenswrapper[4953]: I1203 00:12:06.965922 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Dec 03 00:12:06 crc kubenswrapper[4953]: I1203 00:12:06.990855 4953 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Dec 03 00:12:06 crc kubenswrapper[4953]: I1203 00:12:06.994459 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Dec 03 00:12:07 crc kubenswrapper[4953]: I1203 00:12:07.106691 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Dec 03 00:12:07 crc kubenswrapper[4953]: I1203 00:12:07.225329 4953 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Dec 03 00:12:07 crc kubenswrapper[4953]: I1203 00:12:07.229209 4953 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Dec 03 00:12:07 crc kubenswrapper[4953]: I1203 00:12:07.371285 4953 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Dec 03 00:12:07 crc kubenswrapper[4953]: I1203 00:12:07.375021 4953 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Dec 03 00:12:07 crc kubenswrapper[4953]: I1203 00:12:07.424555 4953 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Dec 03 00:12:07 crc kubenswrapper[4953]: I1203 00:12:07.487327 4953 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Dec 03 00:12:07 crc kubenswrapper[4953]: I1203 00:12:07.528443 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Dec 03 00:12:07 crc kubenswrapper[4953]: I1203 00:12:07.538441 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Dec 03 00:12:07 crc kubenswrapper[4953]: I1203 00:12:07.552465 4953 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Dec 03 00:12:07 crc kubenswrapper[4953]: I1203 00:12:07.609815 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Dec 03 00:12:07 crc kubenswrapper[4953]: I1203 00:12:07.635378 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Dec 03 00:12:07 crc kubenswrapper[4953]: I1203 00:12:07.693826 4953 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Dec 03 00:12:07 crc kubenswrapper[4953]: I1203 00:12:07.741263 4953 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Dec 03 00:12:07 crc kubenswrapper[4953]: I1203 00:12:07.858022 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Dec 03 00:12:07 crc kubenswrapper[4953]: I1203 00:12:07.904334 4953 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Dec 03 00:12:07 crc kubenswrapper[4953]: I1203 00:12:07.909630 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Dec 03 00:12:08 crc kubenswrapper[4953]: I1203 00:12:08.037006 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Dec 03 00:12:08 crc kubenswrapper[4953]: I1203 00:12:08.159306 4953 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Dec 03 00:12:08 crc kubenswrapper[4953]: I1203 00:12:08.160125 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Dec 03 00:12:08 crc kubenswrapper[4953]: I1203 00:12:08.347416 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Dec 03 00:12:08 crc kubenswrapper[4953]: I1203 00:12:08.368906 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Dec 03 00:12:08 crc kubenswrapper[4953]: I1203 00:12:08.484763 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Dec 03 00:12:08 crc kubenswrapper[4953]: I1203 00:12:08.533237 4953 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Dec 03 00:12:08 crc kubenswrapper[4953]: I1203 00:12:08.537520 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Dec 03 00:12:08 crc kubenswrapper[4953]: I1203 00:12:08.573023 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Dec 03 00:12:08 crc kubenswrapper[4953]: I1203 00:12:08.590270 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Dec 03 00:12:08 crc kubenswrapper[4953]: I1203 00:12:08.631603 4953 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Dec 03 00:12:08 crc kubenswrapper[4953]: I1203 00:12:08.694014 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Dec 03 00:12:08 crc kubenswrapper[4953]: I1203 00:12:08.696196 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Dec 03 00:12:08 crc kubenswrapper[4953]: I1203 00:12:08.734427 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Dec 03 00:12:08 crc kubenswrapper[4953]: I1203 00:12:08.762763 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Dec 03 00:12:08 crc kubenswrapper[4953]: I1203 00:12:08.835044 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Dec 03 00:12:08 crc kubenswrapper[4953]: I1203 00:12:08.857821 4953 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Dec 03 00:12:08 crc kubenswrapper[4953]: I1203 00:12:08.868869 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Dec 03 00:12:08 crc kubenswrapper[4953]: I1203 00:12:08.897145 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Dec 03 00:12:08 crc kubenswrapper[4953]: I1203 00:12:08.913602 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Dec 03 00:12:09 crc kubenswrapper[4953]: I1203 00:12:09.033721 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Dec 03 00:12:09 crc kubenswrapper[4953]: I1203 00:12:09.033820 4953 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Dec 03 00:12:09 crc kubenswrapper[4953]: I1203 00:12:09.061989 4953 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Dec 03 00:12:09 crc kubenswrapper[4953]: I1203 00:12:09.150746 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Dec 03 00:12:09 crc kubenswrapper[4953]: I1203 00:12:09.182448 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Dec 03 00:12:09 crc kubenswrapper[4953]: I1203 00:12:09.198253 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Dec 03 00:12:09 crc kubenswrapper[4953]: I1203 00:12:09.307316 4953 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Dec 03 00:12:09 crc kubenswrapper[4953]: I1203 00:12:09.447843 4953 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Dec 03 00:12:09 crc kubenswrapper[4953]: I1203 00:12:09.518571 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Dec 03 00:12:09 crc kubenswrapper[4953]: I1203 00:12:09.538884 4953 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Dec 03 00:12:09 crc kubenswrapper[4953]: I1203 00:12:09.619794 4953 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Dec 03 00:12:09 crc kubenswrapper[4953]: I1203 00:12:09.742346 4953 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Dec 03 00:12:09 crc kubenswrapper[4953]: I1203 00:12:09.766937 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Dec 03 00:12:09 crc kubenswrapper[4953]: I1203 00:12:09.815393 4953 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Dec 03 00:12:09 crc kubenswrapper[4953]: I1203 00:12:09.828635 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Dec 03 00:12:09 crc kubenswrapper[4953]: I1203 00:12:09.892876 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Dec 03 00:12:09 crc kubenswrapper[4953]: I1203 00:12:09.990451 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Dec 03 00:12:09 crc kubenswrapper[4953]: I1203 00:12:09.990487 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Dec 03 00:12:10 crc kubenswrapper[4953]: I1203 00:12:10.096388 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Dec 03 00:12:10 crc kubenswrapper[4953]: I1203 00:12:10.181532 4953 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Dec 03 00:12:10 crc kubenswrapper[4953]: I1203 00:12:10.210633 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Dec 03 00:12:10 crc kubenswrapper[4953]: I1203 00:12:10.344790 4953 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Dec 03 00:12:10 crc kubenswrapper[4953]: I1203 00:12:10.443343 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Dec 03 00:12:10 crc kubenswrapper[4953]: I1203 00:12:10.488848 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Dec 03 00:12:10 crc kubenswrapper[4953]: I1203 00:12:10.515892 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Dec 03 00:12:10 crc kubenswrapper[4953]: I1203 00:12:10.540155 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Dec 03 00:12:10 crc kubenswrapper[4953]: I1203 00:12:10.541306 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Dec 03 00:12:10 crc kubenswrapper[4953]: I1203 00:12:10.674422 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Dec 03 00:12:10 crc kubenswrapper[4953]: I1203 00:12:10.762930 4953 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Dec 03 00:12:10 crc kubenswrapper[4953]: I1203 00:12:10.824900 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Dec 03 00:12:10 crc kubenswrapper[4953]: I1203 00:12:10.840732 4953 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Dec 03 00:12:10 crc kubenswrapper[4953]: I1203 00:12:10.846899 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Dec 03 00:12:10 crc kubenswrapper[4953]: I1203 00:12:10.877730 4953 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Dec 03 00:12:10 crc kubenswrapper[4953]: I1203 00:12:10.934355 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Dec 03 00:12:11 crc kubenswrapper[4953]: I1203 00:12:11.100961 4953 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Dec 03 00:12:11 crc kubenswrapper[4953]: I1203 00:12:11.140510 4953 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Dec 03 00:12:11 crc kubenswrapper[4953]: I1203 00:12:11.154998 4953 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Dec 03 00:12:11 crc kubenswrapper[4953]: I1203 00:12:11.193888 4953 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Dec 03 00:12:11 crc kubenswrapper[4953]: I1203 00:12:11.206469 4953 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Dec 03 00:12:11 crc kubenswrapper[4953]: I1203 00:12:11.253790 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Dec 03 00:12:11 crc kubenswrapper[4953]: I1203 00:12:11.285434 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Dec 03 00:12:11 crc kubenswrapper[4953]: I1203 00:12:11.331755 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Dec 03 00:12:11 crc kubenswrapper[4953]: I1203 00:12:11.345958 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Dec 03 00:12:11 crc kubenswrapper[4953]: I1203 00:12:11.407720 4953 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Dec 03 00:12:11 crc kubenswrapper[4953]: I1203 00:12:11.452483 4953 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Dec 03 00:12:11 crc kubenswrapper[4953]: I1203 00:12:11.535173 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Dec 03 00:12:11 crc kubenswrapper[4953]: I1203 00:12:11.609502 4953 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Dec 03 00:12:11 crc kubenswrapper[4953]: I1203 00:12:11.609888 4953 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" containerID="cri-o://c09cfbc07723f81e802a591039a981ac41ceca4a53e918b96d26ec5f5a633ba8" gracePeriod=5 Dec 03 00:12:11 crc kubenswrapper[4953]: I1203 00:12:11.635641 4953 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-7765894ccc-262rn"] Dec 03 00:12:11 crc kubenswrapper[4953]: E1203 00:12:11.635928 4953 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ca879198-0a45-4505-b845-60d8b0f5bb5d" containerName="oauth-openshift" Dec 03 00:12:11 crc kubenswrapper[4953]: I1203 00:12:11.635945 4953 state_mem.go:107] "Deleted CPUSet assignment" podUID="ca879198-0a45-4505-b845-60d8b0f5bb5d" containerName="oauth-openshift" Dec 03 00:12:11 crc kubenswrapper[4953]: E1203 00:12:11.635963 4953 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bc833625-0910-40d3-8f2d-801b4b9c40d2" containerName="installer" Dec 03 00:12:11 crc kubenswrapper[4953]: I1203 00:12:11.635988 4953 state_mem.go:107] "Deleted CPUSet assignment" podUID="bc833625-0910-40d3-8f2d-801b4b9c40d2" containerName="installer" Dec 03 00:12:11 crc kubenswrapper[4953]: E1203 00:12:11.636005 4953 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Dec 03 00:12:11 crc kubenswrapper[4953]: I1203 00:12:11.636012 4953 state_mem.go:107] "Deleted CPUSet assignment" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Dec 03 00:12:11 crc kubenswrapper[4953]: I1203 00:12:11.636134 4953 memory_manager.go:354] "RemoveStaleState removing state" podUID="bc833625-0910-40d3-8f2d-801b4b9c40d2" containerName="installer" Dec 03 00:12:11 crc kubenswrapper[4953]: I1203 00:12:11.636143 4953 memory_manager.go:354] "RemoveStaleState removing state" podUID="ca879198-0a45-4505-b845-60d8b0f5bb5d" containerName="oauth-openshift" Dec 03 00:12:11 crc kubenswrapper[4953]: I1203 00:12:11.636157 4953 memory_manager.go:354] "RemoveStaleState removing state" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Dec 03 00:12:11 crc kubenswrapper[4953]: I1203 00:12:11.636738 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-7765894ccc-262rn" Dec 03 00:12:11 crc kubenswrapper[4953]: I1203 00:12:11.639677 4953 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Dec 03 00:12:11 crc kubenswrapper[4953]: I1203 00:12:11.639819 4953 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Dec 03 00:12:11 crc kubenswrapper[4953]: I1203 00:12:11.639951 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Dec 03 00:12:11 crc kubenswrapper[4953]: I1203 00:12:11.640119 4953 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Dec 03 00:12:11 crc kubenswrapper[4953]: I1203 00:12:11.640308 4953 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Dec 03 00:12:11 crc kubenswrapper[4953]: I1203 00:12:11.640410 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Dec 03 00:12:11 crc kubenswrapper[4953]: I1203 00:12:11.642528 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Dec 03 00:12:11 crc kubenswrapper[4953]: I1203 00:12:11.643110 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Dec 03 00:12:11 crc kubenswrapper[4953]: I1203 00:12:11.645803 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Dec 03 00:12:11 crc kubenswrapper[4953]: I1203 00:12:11.645929 4953 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Dec 03 00:12:11 crc kubenswrapper[4953]: I1203 00:12:11.645967 4953 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Dec 03 00:12:11 crc kubenswrapper[4953]: I1203 00:12:11.646750 4953 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Dec 03 00:12:11 crc kubenswrapper[4953]: I1203 00:12:11.650370 4953 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Dec 03 00:12:11 crc kubenswrapper[4953]: I1203 00:12:11.654154 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Dec 03 00:12:11 crc kubenswrapper[4953]: I1203 00:12:11.657457 4953 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Dec 03 00:12:11 crc kubenswrapper[4953]: I1203 00:12:11.724647 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Dec 03 00:12:11 crc kubenswrapper[4953]: I1203 00:12:11.732706 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/2014e4ab-7b1f-47f6-aa19-98636c47328a-v4-0-config-system-session\") pod \"oauth-openshift-7765894ccc-262rn\" (UID: \"2014e4ab-7b1f-47f6-aa19-98636c47328a\") " pod="openshift-authentication/oauth-openshift-7765894ccc-262rn" Dec 03 00:12:11 crc kubenswrapper[4953]: I1203 00:12:11.732885 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2014e4ab-7b1f-47f6-aa19-98636c47328a-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-7765894ccc-262rn\" (UID: \"2014e4ab-7b1f-47f6-aa19-98636c47328a\") " pod="openshift-authentication/oauth-openshift-7765894ccc-262rn" Dec 03 00:12:11 crc kubenswrapper[4953]: I1203 00:12:11.732954 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/2014e4ab-7b1f-47f6-aa19-98636c47328a-v4-0-config-user-template-login\") pod \"oauth-openshift-7765894ccc-262rn\" (UID: \"2014e4ab-7b1f-47f6-aa19-98636c47328a\") " pod="openshift-authentication/oauth-openshift-7765894ccc-262rn" Dec 03 00:12:11 crc kubenswrapper[4953]: I1203 00:12:11.733054 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/2014e4ab-7b1f-47f6-aa19-98636c47328a-v4-0-config-system-serving-cert\") pod \"oauth-openshift-7765894ccc-262rn\" (UID: \"2014e4ab-7b1f-47f6-aa19-98636c47328a\") " pod="openshift-authentication/oauth-openshift-7765894ccc-262rn" Dec 03 00:12:11 crc kubenswrapper[4953]: I1203 00:12:11.733133 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/2014e4ab-7b1f-47f6-aa19-98636c47328a-audit-policies\") pod \"oauth-openshift-7765894ccc-262rn\" (UID: \"2014e4ab-7b1f-47f6-aa19-98636c47328a\") " pod="openshift-authentication/oauth-openshift-7765894ccc-262rn" Dec 03 00:12:11 crc kubenswrapper[4953]: I1203 00:12:11.733203 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/2014e4ab-7b1f-47f6-aa19-98636c47328a-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-7765894ccc-262rn\" (UID: \"2014e4ab-7b1f-47f6-aa19-98636c47328a\") " pod="openshift-authentication/oauth-openshift-7765894ccc-262rn" Dec 03 00:12:11 crc kubenswrapper[4953]: I1203 00:12:11.733280 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/2014e4ab-7b1f-47f6-aa19-98636c47328a-v4-0-config-system-router-certs\") pod \"oauth-openshift-7765894ccc-262rn\" (UID: \"2014e4ab-7b1f-47f6-aa19-98636c47328a\") " pod="openshift-authentication/oauth-openshift-7765894ccc-262rn" Dec 03 00:12:11 crc kubenswrapper[4953]: I1203 00:12:11.733368 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/2014e4ab-7b1f-47f6-aa19-98636c47328a-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-7765894ccc-262rn\" (UID: \"2014e4ab-7b1f-47f6-aa19-98636c47328a\") " pod="openshift-authentication/oauth-openshift-7765894ccc-262rn" Dec 03 00:12:11 crc kubenswrapper[4953]: I1203 00:12:11.733448 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/2014e4ab-7b1f-47f6-aa19-98636c47328a-v4-0-config-system-service-ca\") pod \"oauth-openshift-7765894ccc-262rn\" (UID: \"2014e4ab-7b1f-47f6-aa19-98636c47328a\") " pod="openshift-authentication/oauth-openshift-7765894ccc-262rn" Dec 03 00:12:11 crc kubenswrapper[4953]: I1203 00:12:11.733636 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/2014e4ab-7b1f-47f6-aa19-98636c47328a-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-7765894ccc-262rn\" (UID: \"2014e4ab-7b1f-47f6-aa19-98636c47328a\") " pod="openshift-authentication/oauth-openshift-7765894ccc-262rn" Dec 03 00:12:11 crc kubenswrapper[4953]: I1203 00:12:11.733742 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/2014e4ab-7b1f-47f6-aa19-98636c47328a-v4-0-config-user-template-error\") pod \"oauth-openshift-7765894ccc-262rn\" (UID: \"2014e4ab-7b1f-47f6-aa19-98636c47328a\") " pod="openshift-authentication/oauth-openshift-7765894ccc-262rn" Dec 03 00:12:11 crc kubenswrapper[4953]: I1203 00:12:11.733830 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/2014e4ab-7b1f-47f6-aa19-98636c47328a-audit-dir\") pod \"oauth-openshift-7765894ccc-262rn\" (UID: \"2014e4ab-7b1f-47f6-aa19-98636c47328a\") " pod="openshift-authentication/oauth-openshift-7765894ccc-262rn" Dec 03 00:12:11 crc kubenswrapper[4953]: I1203 00:12:11.733914 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n89ts\" (UniqueName: \"kubernetes.io/projected/2014e4ab-7b1f-47f6-aa19-98636c47328a-kube-api-access-n89ts\") pod \"oauth-openshift-7765894ccc-262rn\" (UID: \"2014e4ab-7b1f-47f6-aa19-98636c47328a\") " pod="openshift-authentication/oauth-openshift-7765894ccc-262rn" Dec 03 00:12:11 crc kubenswrapper[4953]: I1203 00:12:11.733959 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/2014e4ab-7b1f-47f6-aa19-98636c47328a-v4-0-config-system-cliconfig\") pod \"oauth-openshift-7765894ccc-262rn\" (UID: \"2014e4ab-7b1f-47f6-aa19-98636c47328a\") " pod="openshift-authentication/oauth-openshift-7765894ccc-262rn" Dec 03 00:12:11 crc kubenswrapper[4953]: I1203 00:12:11.812524 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Dec 03 00:12:11 crc kubenswrapper[4953]: I1203 00:12:11.819240 4953 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Dec 03 00:12:11 crc kubenswrapper[4953]: I1203 00:12:11.836739 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2014e4ab-7b1f-47f6-aa19-98636c47328a-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-7765894ccc-262rn\" (UID: \"2014e4ab-7b1f-47f6-aa19-98636c47328a\") " pod="openshift-authentication/oauth-openshift-7765894ccc-262rn" Dec 03 00:12:11 crc kubenswrapper[4953]: I1203 00:12:11.836801 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/2014e4ab-7b1f-47f6-aa19-98636c47328a-v4-0-config-user-template-login\") pod \"oauth-openshift-7765894ccc-262rn\" (UID: \"2014e4ab-7b1f-47f6-aa19-98636c47328a\") " pod="openshift-authentication/oauth-openshift-7765894ccc-262rn" Dec 03 00:12:11 crc kubenswrapper[4953]: I1203 00:12:11.836838 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/2014e4ab-7b1f-47f6-aa19-98636c47328a-v4-0-config-system-serving-cert\") pod \"oauth-openshift-7765894ccc-262rn\" (UID: \"2014e4ab-7b1f-47f6-aa19-98636c47328a\") " pod="openshift-authentication/oauth-openshift-7765894ccc-262rn" Dec 03 00:12:11 crc kubenswrapper[4953]: I1203 00:12:11.836870 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/2014e4ab-7b1f-47f6-aa19-98636c47328a-audit-policies\") pod \"oauth-openshift-7765894ccc-262rn\" (UID: \"2014e4ab-7b1f-47f6-aa19-98636c47328a\") " pod="openshift-authentication/oauth-openshift-7765894ccc-262rn" Dec 03 00:12:11 crc kubenswrapper[4953]: I1203 00:12:11.836917 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/2014e4ab-7b1f-47f6-aa19-98636c47328a-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-7765894ccc-262rn\" (UID: \"2014e4ab-7b1f-47f6-aa19-98636c47328a\") " pod="openshift-authentication/oauth-openshift-7765894ccc-262rn" Dec 03 00:12:11 crc kubenswrapper[4953]: I1203 00:12:11.836953 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/2014e4ab-7b1f-47f6-aa19-98636c47328a-v4-0-config-system-router-certs\") pod \"oauth-openshift-7765894ccc-262rn\" (UID: \"2014e4ab-7b1f-47f6-aa19-98636c47328a\") " pod="openshift-authentication/oauth-openshift-7765894ccc-262rn" Dec 03 00:12:11 crc kubenswrapper[4953]: I1203 00:12:11.837001 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/2014e4ab-7b1f-47f6-aa19-98636c47328a-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-7765894ccc-262rn\" (UID: \"2014e4ab-7b1f-47f6-aa19-98636c47328a\") " pod="openshift-authentication/oauth-openshift-7765894ccc-262rn" Dec 03 00:12:11 crc kubenswrapper[4953]: I1203 00:12:11.837043 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/2014e4ab-7b1f-47f6-aa19-98636c47328a-v4-0-config-system-service-ca\") pod \"oauth-openshift-7765894ccc-262rn\" (UID: \"2014e4ab-7b1f-47f6-aa19-98636c47328a\") " pod="openshift-authentication/oauth-openshift-7765894ccc-262rn" Dec 03 00:12:11 crc kubenswrapper[4953]: I1203 00:12:11.837084 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/2014e4ab-7b1f-47f6-aa19-98636c47328a-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-7765894ccc-262rn\" (UID: \"2014e4ab-7b1f-47f6-aa19-98636c47328a\") " pod="openshift-authentication/oauth-openshift-7765894ccc-262rn" Dec 03 00:12:11 crc kubenswrapper[4953]: I1203 00:12:11.837117 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/2014e4ab-7b1f-47f6-aa19-98636c47328a-v4-0-config-user-template-error\") pod \"oauth-openshift-7765894ccc-262rn\" (UID: \"2014e4ab-7b1f-47f6-aa19-98636c47328a\") " pod="openshift-authentication/oauth-openshift-7765894ccc-262rn" Dec 03 00:12:11 crc kubenswrapper[4953]: I1203 00:12:11.837143 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n89ts\" (UniqueName: \"kubernetes.io/projected/2014e4ab-7b1f-47f6-aa19-98636c47328a-kube-api-access-n89ts\") pod \"oauth-openshift-7765894ccc-262rn\" (UID: \"2014e4ab-7b1f-47f6-aa19-98636c47328a\") " pod="openshift-authentication/oauth-openshift-7765894ccc-262rn" Dec 03 00:12:11 crc kubenswrapper[4953]: I1203 00:12:11.837167 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/2014e4ab-7b1f-47f6-aa19-98636c47328a-audit-dir\") pod \"oauth-openshift-7765894ccc-262rn\" (UID: \"2014e4ab-7b1f-47f6-aa19-98636c47328a\") " pod="openshift-authentication/oauth-openshift-7765894ccc-262rn" Dec 03 00:12:11 crc kubenswrapper[4953]: I1203 00:12:11.837193 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/2014e4ab-7b1f-47f6-aa19-98636c47328a-v4-0-config-system-cliconfig\") pod \"oauth-openshift-7765894ccc-262rn\" (UID: \"2014e4ab-7b1f-47f6-aa19-98636c47328a\") " pod="openshift-authentication/oauth-openshift-7765894ccc-262rn" Dec 03 00:12:11 crc kubenswrapper[4953]: I1203 00:12:11.837227 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/2014e4ab-7b1f-47f6-aa19-98636c47328a-v4-0-config-system-session\") pod \"oauth-openshift-7765894ccc-262rn\" (UID: \"2014e4ab-7b1f-47f6-aa19-98636c47328a\") " pod="openshift-authentication/oauth-openshift-7765894ccc-262rn" Dec 03 00:12:11 crc kubenswrapper[4953]: I1203 00:12:11.838336 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/2014e4ab-7b1f-47f6-aa19-98636c47328a-audit-dir\") pod \"oauth-openshift-7765894ccc-262rn\" (UID: \"2014e4ab-7b1f-47f6-aa19-98636c47328a\") " pod="openshift-authentication/oauth-openshift-7765894ccc-262rn" Dec 03 00:12:11 crc kubenswrapper[4953]: I1203 00:12:11.838617 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/2014e4ab-7b1f-47f6-aa19-98636c47328a-v4-0-config-system-service-ca\") pod \"oauth-openshift-7765894ccc-262rn\" (UID: \"2014e4ab-7b1f-47f6-aa19-98636c47328a\") " pod="openshift-authentication/oauth-openshift-7765894ccc-262rn" Dec 03 00:12:11 crc kubenswrapper[4953]: I1203 00:12:11.839383 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/2014e4ab-7b1f-47f6-aa19-98636c47328a-audit-policies\") pod \"oauth-openshift-7765894ccc-262rn\" (UID: \"2014e4ab-7b1f-47f6-aa19-98636c47328a\") " pod="openshift-authentication/oauth-openshift-7765894ccc-262rn" Dec 03 00:12:11 crc kubenswrapper[4953]: I1203 00:12:11.839448 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/2014e4ab-7b1f-47f6-aa19-98636c47328a-v4-0-config-system-cliconfig\") pod \"oauth-openshift-7765894ccc-262rn\" (UID: \"2014e4ab-7b1f-47f6-aa19-98636c47328a\") " pod="openshift-authentication/oauth-openshift-7765894ccc-262rn" Dec 03 00:12:11 crc kubenswrapper[4953]: I1203 00:12:11.839592 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2014e4ab-7b1f-47f6-aa19-98636c47328a-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-7765894ccc-262rn\" (UID: \"2014e4ab-7b1f-47f6-aa19-98636c47328a\") " pod="openshift-authentication/oauth-openshift-7765894ccc-262rn" Dec 03 00:12:11 crc kubenswrapper[4953]: I1203 00:12:11.843332 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Dec 03 00:12:11 crc kubenswrapper[4953]: I1203 00:12:11.845653 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/2014e4ab-7b1f-47f6-aa19-98636c47328a-v4-0-config-system-serving-cert\") pod \"oauth-openshift-7765894ccc-262rn\" (UID: \"2014e4ab-7b1f-47f6-aa19-98636c47328a\") " pod="openshift-authentication/oauth-openshift-7765894ccc-262rn" Dec 03 00:12:11 crc kubenswrapper[4953]: I1203 00:12:11.845794 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/2014e4ab-7b1f-47f6-aa19-98636c47328a-v4-0-config-system-router-certs\") pod \"oauth-openshift-7765894ccc-262rn\" (UID: \"2014e4ab-7b1f-47f6-aa19-98636c47328a\") " pod="openshift-authentication/oauth-openshift-7765894ccc-262rn" Dec 03 00:12:11 crc kubenswrapper[4953]: I1203 00:12:11.846385 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/2014e4ab-7b1f-47f6-aa19-98636c47328a-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-7765894ccc-262rn\" (UID: \"2014e4ab-7b1f-47f6-aa19-98636c47328a\") " pod="openshift-authentication/oauth-openshift-7765894ccc-262rn" Dec 03 00:12:11 crc kubenswrapper[4953]: I1203 00:12:11.846546 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/2014e4ab-7b1f-47f6-aa19-98636c47328a-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-7765894ccc-262rn\" (UID: \"2014e4ab-7b1f-47f6-aa19-98636c47328a\") " pod="openshift-authentication/oauth-openshift-7765894ccc-262rn" Dec 03 00:12:11 crc kubenswrapper[4953]: I1203 00:12:11.847102 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/2014e4ab-7b1f-47f6-aa19-98636c47328a-v4-0-config-user-template-login\") pod \"oauth-openshift-7765894ccc-262rn\" (UID: \"2014e4ab-7b1f-47f6-aa19-98636c47328a\") " pod="openshift-authentication/oauth-openshift-7765894ccc-262rn" Dec 03 00:12:11 crc kubenswrapper[4953]: I1203 00:12:11.847125 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/2014e4ab-7b1f-47f6-aa19-98636c47328a-v4-0-config-user-template-error\") pod \"oauth-openshift-7765894ccc-262rn\" (UID: \"2014e4ab-7b1f-47f6-aa19-98636c47328a\") " pod="openshift-authentication/oauth-openshift-7765894ccc-262rn" Dec 03 00:12:11 crc kubenswrapper[4953]: I1203 00:12:11.852690 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Dec 03 00:12:11 crc kubenswrapper[4953]: I1203 00:12:11.852947 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/2014e4ab-7b1f-47f6-aa19-98636c47328a-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-7765894ccc-262rn\" (UID: \"2014e4ab-7b1f-47f6-aa19-98636c47328a\") " pod="openshift-authentication/oauth-openshift-7765894ccc-262rn" Dec 03 00:12:11 crc kubenswrapper[4953]: I1203 00:12:11.859838 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n89ts\" (UniqueName: \"kubernetes.io/projected/2014e4ab-7b1f-47f6-aa19-98636c47328a-kube-api-access-n89ts\") pod \"oauth-openshift-7765894ccc-262rn\" (UID: \"2014e4ab-7b1f-47f6-aa19-98636c47328a\") " pod="openshift-authentication/oauth-openshift-7765894ccc-262rn" Dec 03 00:12:11 crc kubenswrapper[4953]: I1203 00:12:11.863289 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/2014e4ab-7b1f-47f6-aa19-98636c47328a-v4-0-config-system-session\") pod \"oauth-openshift-7765894ccc-262rn\" (UID: \"2014e4ab-7b1f-47f6-aa19-98636c47328a\") " pod="openshift-authentication/oauth-openshift-7765894ccc-262rn" Dec 03 00:12:11 crc kubenswrapper[4953]: I1203 00:12:11.892416 4953 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Dec 03 00:12:11 crc kubenswrapper[4953]: I1203 00:12:11.907828 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Dec 03 00:12:11 crc kubenswrapper[4953]: I1203 00:12:11.968809 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-7765894ccc-262rn" Dec 03 00:12:11 crc kubenswrapper[4953]: I1203 00:12:11.986074 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Dec 03 00:12:12 crc kubenswrapper[4953]: I1203 00:12:12.149967 4953 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Dec 03 00:12:12 crc kubenswrapper[4953]: I1203 00:12:12.173833 4953 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Dec 03 00:12:12 crc kubenswrapper[4953]: I1203 00:12:12.186793 4953 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Dec 03 00:12:12 crc kubenswrapper[4953]: I1203 00:12:12.265144 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Dec 03 00:12:12 crc kubenswrapper[4953]: I1203 00:12:12.615964 4953 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Dec 03 00:12:12 crc kubenswrapper[4953]: I1203 00:12:12.667718 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Dec 03 00:12:12 crc kubenswrapper[4953]: I1203 00:12:12.840267 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Dec 03 00:12:12 crc kubenswrapper[4953]: I1203 00:12:12.886157 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Dec 03 00:12:12 crc kubenswrapper[4953]: I1203 00:12:12.888364 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Dec 03 00:12:12 crc kubenswrapper[4953]: I1203 00:12:12.924338 4953 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Dec 03 00:12:12 crc kubenswrapper[4953]: I1203 00:12:12.951806 4953 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Dec 03 00:12:13 crc kubenswrapper[4953]: I1203 00:12:13.297400 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Dec 03 00:12:13 crc kubenswrapper[4953]: I1203 00:12:13.313716 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Dec 03 00:12:13 crc kubenswrapper[4953]: I1203 00:12:13.381452 4953 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Dec 03 00:12:13 crc kubenswrapper[4953]: I1203 00:12:13.531280 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Dec 03 00:12:13 crc kubenswrapper[4953]: I1203 00:12:13.629824 4953 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-7765894ccc-262rn"] Dec 03 00:12:13 crc kubenswrapper[4953]: I1203 00:12:13.706999 4953 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Dec 03 00:12:13 crc kubenswrapper[4953]: I1203 00:12:13.798300 4953 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Dec 03 00:12:13 crc kubenswrapper[4953]: I1203 00:12:13.829403 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Dec 03 00:12:13 crc kubenswrapper[4953]: I1203 00:12:13.830673 4953 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-7765894ccc-262rn"] Dec 03 00:12:14 crc kubenswrapper[4953]: I1203 00:12:14.163030 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Dec 03 00:12:14 crc kubenswrapper[4953]: I1203 00:12:14.295959 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-7765894ccc-262rn" event={"ID":"2014e4ab-7b1f-47f6-aa19-98636c47328a","Type":"ContainerStarted","Data":"2dda2469db8952133f543452200593995f4ce9368872d234ca7f0684afb8088f"} Dec 03 00:12:14 crc kubenswrapper[4953]: I1203 00:12:14.296069 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-7765894ccc-262rn" event={"ID":"2014e4ab-7b1f-47f6-aa19-98636c47328a","Type":"ContainerStarted","Data":"d7b4647b9c4976bb435379a995bea04e88c88a127b2440cf422fe88d7c2fa286"} Dec 03 00:12:14 crc kubenswrapper[4953]: I1203 00:12:14.297449 4953 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-7765894ccc-262rn" Dec 03 00:12:14 crc kubenswrapper[4953]: I1203 00:12:14.318088 4953 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-7765894ccc-262rn" podStartSLOduration=56.318063229 podStartE2EDuration="56.318063229s" podCreationTimestamp="2025-12-03 00:11:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 00:12:14.315305129 +0000 UTC m=+323.299585912" watchObservedRunningTime="2025-12-03 00:12:14.318063229 +0000 UTC m=+323.302344012" Dec 03 00:12:14 crc kubenswrapper[4953]: I1203 00:12:14.394155 4953 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Dec 03 00:12:14 crc kubenswrapper[4953]: I1203 00:12:14.669095 4953 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Dec 03 00:12:14 crc kubenswrapper[4953]: I1203 00:12:14.740022 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Dec 03 00:12:14 crc kubenswrapper[4953]: I1203 00:12:14.768372 4953 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Dec 03 00:12:14 crc kubenswrapper[4953]: I1203 00:12:14.942509 4953 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Dec 03 00:12:14 crc kubenswrapper[4953]: I1203 00:12:14.965087 4953 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Dec 03 00:12:15 crc kubenswrapper[4953]: I1203 00:12:15.020489 4953 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-7765894ccc-262rn" Dec 03 00:12:15 crc kubenswrapper[4953]: I1203 00:12:15.327178 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Dec 03 00:12:16 crc kubenswrapper[4953]: I1203 00:12:16.422510 4953 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Dec 03 00:12:17 crc kubenswrapper[4953]: I1203 00:12:17.222439 4953 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Dec 03 00:12:17 crc kubenswrapper[4953]: I1203 00:12:17.222785 4953 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 03 00:12:17 crc kubenswrapper[4953]: I1203 00:12:17.315051 4953 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Dec 03 00:12:17 crc kubenswrapper[4953]: I1203 00:12:17.315120 4953 generic.go:334] "Generic (PLEG): container finished" podID="f85e55b1a89d02b0cb034b1ea31ed45a" containerID="c09cfbc07723f81e802a591039a981ac41ceca4a53e918b96d26ec5f5a633ba8" exitCode=137 Dec 03 00:12:17 crc kubenswrapper[4953]: I1203 00:12:17.315204 4953 scope.go:117] "RemoveContainer" containerID="c09cfbc07723f81e802a591039a981ac41ceca4a53e918b96d26ec5f5a633ba8" Dec 03 00:12:17 crc kubenswrapper[4953]: I1203 00:12:17.315211 4953 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 03 00:12:17 crc kubenswrapper[4953]: I1203 00:12:17.327631 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 03 00:12:17 crc kubenswrapper[4953]: I1203 00:12:17.327710 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 03 00:12:17 crc kubenswrapper[4953]: I1203 00:12:17.327798 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 03 00:12:17 crc kubenswrapper[4953]: I1203 00:12:17.327819 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 03 00:12:17 crc kubenswrapper[4953]: I1203 00:12:17.327845 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 03 00:12:17 crc kubenswrapper[4953]: I1203 00:12:17.327917 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 00:12:17 crc kubenswrapper[4953]: I1203 00:12:17.327904 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock" (OuterVolumeSpecName: "var-lock") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 00:12:17 crc kubenswrapper[4953]: I1203 00:12:17.328068 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests" (OuterVolumeSpecName: "manifests") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "manifests". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 00:12:17 crc kubenswrapper[4953]: I1203 00:12:17.328082 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log" (OuterVolumeSpecName: "var-log") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 00:12:17 crc kubenswrapper[4953]: I1203 00:12:17.328566 4953 reconciler_common.go:293] "Volume detached for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") on node \"crc\" DevicePath \"\"" Dec 03 00:12:17 crc kubenswrapper[4953]: I1203 00:12:17.328617 4953 reconciler_common.go:293] "Volume detached for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") on node \"crc\" DevicePath \"\"" Dec 03 00:12:17 crc kubenswrapper[4953]: I1203 00:12:17.328644 4953 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") on node \"crc\" DevicePath \"\"" Dec 03 00:12:17 crc kubenswrapper[4953]: I1203 00:12:17.328671 4953 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") on node \"crc\" DevicePath \"\"" Dec 03 00:12:17 crc kubenswrapper[4953]: I1203 00:12:17.336812 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir" (OuterVolumeSpecName: "pod-resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "pod-resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 00:12:17 crc kubenswrapper[4953]: I1203 00:12:17.345421 4953 scope.go:117] "RemoveContainer" containerID="c09cfbc07723f81e802a591039a981ac41ceca4a53e918b96d26ec5f5a633ba8" Dec 03 00:12:17 crc kubenswrapper[4953]: E1203 00:12:17.346225 4953 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c09cfbc07723f81e802a591039a981ac41ceca4a53e918b96d26ec5f5a633ba8\": container with ID starting with c09cfbc07723f81e802a591039a981ac41ceca4a53e918b96d26ec5f5a633ba8 not found: ID does not exist" containerID="c09cfbc07723f81e802a591039a981ac41ceca4a53e918b96d26ec5f5a633ba8" Dec 03 00:12:17 crc kubenswrapper[4953]: I1203 00:12:17.346272 4953 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c09cfbc07723f81e802a591039a981ac41ceca4a53e918b96d26ec5f5a633ba8"} err="failed to get container status \"c09cfbc07723f81e802a591039a981ac41ceca4a53e918b96d26ec5f5a633ba8\": rpc error: code = NotFound desc = could not find container \"c09cfbc07723f81e802a591039a981ac41ceca4a53e918b96d26ec5f5a633ba8\": container with ID starting with c09cfbc07723f81e802a591039a981ac41ceca4a53e918b96d26ec5f5a633ba8 not found: ID does not exist" Dec 03 00:12:17 crc kubenswrapper[4953]: I1203 00:12:17.356592 4953 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" path="/var/lib/kubelet/pods/f85e55b1a89d02b0cb034b1ea31ed45a/volumes" Dec 03 00:12:17 crc kubenswrapper[4953]: I1203 00:12:17.430517 4953 reconciler_common.go:293] "Volume detached for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") on node \"crc\" DevicePath \"\"" Dec 03 00:12:25 crc kubenswrapper[4953]: I1203 00:12:25.969799 4953 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Dec 03 00:12:31 crc kubenswrapper[4953]: I1203 00:12:31.402717 4953 generic.go:334] "Generic (PLEG): container finished" podID="7124b81d-b87c-45de-a112-20602c4d9adb" containerID="cf37583f0764dfdbf86f5e9f0dae7b270a80910c7793261d30e36b161b46cea5" exitCode=0 Dec 03 00:12:31 crc kubenswrapper[4953]: I1203 00:12:31.402798 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-kx9dt" event={"ID":"7124b81d-b87c-45de-a112-20602c4d9adb","Type":"ContainerDied","Data":"cf37583f0764dfdbf86f5e9f0dae7b270a80910c7793261d30e36b161b46cea5"} Dec 03 00:12:31 crc kubenswrapper[4953]: I1203 00:12:31.403730 4953 scope.go:117] "RemoveContainer" containerID="cf37583f0764dfdbf86f5e9f0dae7b270a80910c7793261d30e36b161b46cea5" Dec 03 00:12:32 crc kubenswrapper[4953]: I1203 00:12:32.421357 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-kx9dt" event={"ID":"7124b81d-b87c-45de-a112-20602c4d9adb","Type":"ContainerStarted","Data":"9a1303fc54264d3be3fb57969f4e7934b0ab3a22cc45c5e3c658bd67035f6751"} Dec 03 00:12:32 crc kubenswrapper[4953]: I1203 00:12:32.422131 4953 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-kx9dt" Dec 03 00:12:32 crc kubenswrapper[4953]: I1203 00:12:32.423811 4953 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-kx9dt" Dec 03 00:12:41 crc kubenswrapper[4953]: I1203 00:12:41.805795 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Dec 03 00:12:42 crc kubenswrapper[4953]: I1203 00:12:42.964699 4953 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Dec 03 00:12:46 crc kubenswrapper[4953]: I1203 00:12:46.424262 4953 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Dec 03 00:12:48 crc kubenswrapper[4953]: I1203 00:12:48.944709 4953 patch_prober.go:28] interesting pod/machine-config-daemon-6dfbn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 00:12:48 crc kubenswrapper[4953]: I1203 00:12:48.945156 4953 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-6dfbn" podUID="53ffb1e9-9a2e-4869-80d7-776f4aa7ef8b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 00:12:55 crc kubenswrapper[4953]: I1203 00:12:55.468076 4953 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-l7sj9"] Dec 03 00:12:55 crc kubenswrapper[4953]: I1203 00:12:55.470924 4953 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-879f6c89f-l7sj9" podUID="b33612b4-02d4-4f26-86bb-73a630d1ee58" containerName="controller-manager" containerID="cri-o://2027e38371dd6192fed82e47860af430e69b879668794026dc7f3b6ba304f8e0" gracePeriod=30 Dec 03 00:12:55 crc kubenswrapper[4953]: I1203 00:12:55.571625 4953 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-52hrr"] Dec 03 00:12:55 crc kubenswrapper[4953]: I1203 00:12:55.572432 4953 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-52hrr" podUID="90247294-af78-4d04-8aed-af90f93a034a" containerName="route-controller-manager" containerID="cri-o://5146c8aac8d7e3cb2d34f3059cb8805fa3cfd735cbdd42ef3ea7e72e404d7cce" gracePeriod=30 Dec 03 00:12:56 crc kubenswrapper[4953]: I1203 00:12:56.042725 4953 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-l7sj9" Dec 03 00:12:56 crc kubenswrapper[4953]: I1203 00:12:56.082952 4953 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-52hrr" Dec 03 00:12:56 crc kubenswrapper[4953]: I1203 00:12:56.105347 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/90247294-af78-4d04-8aed-af90f93a034a-config\") pod \"90247294-af78-4d04-8aed-af90f93a034a\" (UID: \"90247294-af78-4d04-8aed-af90f93a034a\") " Dec 03 00:12:56 crc kubenswrapper[4953]: I1203 00:12:56.105406 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/90247294-af78-4d04-8aed-af90f93a034a-client-ca\") pod \"90247294-af78-4d04-8aed-af90f93a034a\" (UID: \"90247294-af78-4d04-8aed-af90f93a034a\") " Dec 03 00:12:56 crc kubenswrapper[4953]: I1203 00:12:56.105451 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fwqw4\" (UniqueName: \"kubernetes.io/projected/90247294-af78-4d04-8aed-af90f93a034a-kube-api-access-fwqw4\") pod \"90247294-af78-4d04-8aed-af90f93a034a\" (UID: \"90247294-af78-4d04-8aed-af90f93a034a\") " Dec 03 00:12:56 crc kubenswrapper[4953]: I1203 00:12:56.105489 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/90247294-af78-4d04-8aed-af90f93a034a-serving-cert\") pod \"90247294-af78-4d04-8aed-af90f93a034a\" (UID: \"90247294-af78-4d04-8aed-af90f93a034a\") " Dec 03 00:12:56 crc kubenswrapper[4953]: I1203 00:12:56.105526 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b33612b4-02d4-4f26-86bb-73a630d1ee58-config\") pod \"b33612b4-02d4-4f26-86bb-73a630d1ee58\" (UID: \"b33612b4-02d4-4f26-86bb-73a630d1ee58\") " Dec 03 00:12:56 crc kubenswrapper[4953]: I1203 00:12:56.105557 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2vwkp\" (UniqueName: \"kubernetes.io/projected/b33612b4-02d4-4f26-86bb-73a630d1ee58-kube-api-access-2vwkp\") pod \"b33612b4-02d4-4f26-86bb-73a630d1ee58\" (UID: \"b33612b4-02d4-4f26-86bb-73a630d1ee58\") " Dec 03 00:12:56 crc kubenswrapper[4953]: I1203 00:12:56.105589 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b33612b4-02d4-4f26-86bb-73a630d1ee58-serving-cert\") pod \"b33612b4-02d4-4f26-86bb-73a630d1ee58\" (UID: \"b33612b4-02d4-4f26-86bb-73a630d1ee58\") " Dec 03 00:12:56 crc kubenswrapper[4953]: I1203 00:12:56.105605 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/b33612b4-02d4-4f26-86bb-73a630d1ee58-proxy-ca-bundles\") pod \"b33612b4-02d4-4f26-86bb-73a630d1ee58\" (UID: \"b33612b4-02d4-4f26-86bb-73a630d1ee58\") " Dec 03 00:12:56 crc kubenswrapper[4953]: I1203 00:12:56.105643 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b33612b4-02d4-4f26-86bb-73a630d1ee58-client-ca\") pod \"b33612b4-02d4-4f26-86bb-73a630d1ee58\" (UID: \"b33612b4-02d4-4f26-86bb-73a630d1ee58\") " Dec 03 00:12:56 crc kubenswrapper[4953]: I1203 00:12:56.107030 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/90247294-af78-4d04-8aed-af90f93a034a-config" (OuterVolumeSpecName: "config") pod "90247294-af78-4d04-8aed-af90f93a034a" (UID: "90247294-af78-4d04-8aed-af90f93a034a"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:12:56 crc kubenswrapper[4953]: I1203 00:12:56.107193 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b33612b4-02d4-4f26-86bb-73a630d1ee58-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "b33612b4-02d4-4f26-86bb-73a630d1ee58" (UID: "b33612b4-02d4-4f26-86bb-73a630d1ee58"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:12:56 crc kubenswrapper[4953]: I1203 00:12:56.107248 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b33612b4-02d4-4f26-86bb-73a630d1ee58-config" (OuterVolumeSpecName: "config") pod "b33612b4-02d4-4f26-86bb-73a630d1ee58" (UID: "b33612b4-02d4-4f26-86bb-73a630d1ee58"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:12:56 crc kubenswrapper[4953]: I1203 00:12:56.107335 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/90247294-af78-4d04-8aed-af90f93a034a-client-ca" (OuterVolumeSpecName: "client-ca") pod "90247294-af78-4d04-8aed-af90f93a034a" (UID: "90247294-af78-4d04-8aed-af90f93a034a"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:12:56 crc kubenswrapper[4953]: I1203 00:12:56.107954 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b33612b4-02d4-4f26-86bb-73a630d1ee58-client-ca" (OuterVolumeSpecName: "client-ca") pod "b33612b4-02d4-4f26-86bb-73a630d1ee58" (UID: "b33612b4-02d4-4f26-86bb-73a630d1ee58"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:12:56 crc kubenswrapper[4953]: I1203 00:12:56.112324 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b33612b4-02d4-4f26-86bb-73a630d1ee58-kube-api-access-2vwkp" (OuterVolumeSpecName: "kube-api-access-2vwkp") pod "b33612b4-02d4-4f26-86bb-73a630d1ee58" (UID: "b33612b4-02d4-4f26-86bb-73a630d1ee58"). InnerVolumeSpecName "kube-api-access-2vwkp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:12:56 crc kubenswrapper[4953]: I1203 00:12:56.112437 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/90247294-af78-4d04-8aed-af90f93a034a-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "90247294-af78-4d04-8aed-af90f93a034a" (UID: "90247294-af78-4d04-8aed-af90f93a034a"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:12:56 crc kubenswrapper[4953]: I1203 00:12:56.113138 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/90247294-af78-4d04-8aed-af90f93a034a-kube-api-access-fwqw4" (OuterVolumeSpecName: "kube-api-access-fwqw4") pod "90247294-af78-4d04-8aed-af90f93a034a" (UID: "90247294-af78-4d04-8aed-af90f93a034a"). InnerVolumeSpecName "kube-api-access-fwqw4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:12:56 crc kubenswrapper[4953]: I1203 00:12:56.113185 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b33612b4-02d4-4f26-86bb-73a630d1ee58-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "b33612b4-02d4-4f26-86bb-73a630d1ee58" (UID: "b33612b4-02d4-4f26-86bb-73a630d1ee58"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:12:56 crc kubenswrapper[4953]: I1203 00:12:56.206448 4953 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2vwkp\" (UniqueName: \"kubernetes.io/projected/b33612b4-02d4-4f26-86bb-73a630d1ee58-kube-api-access-2vwkp\") on node \"crc\" DevicePath \"\"" Dec 03 00:12:56 crc kubenswrapper[4953]: I1203 00:12:56.206491 4953 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b33612b4-02d4-4f26-86bb-73a630d1ee58-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 00:12:56 crc kubenswrapper[4953]: I1203 00:12:56.206500 4953 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/b33612b4-02d4-4f26-86bb-73a630d1ee58-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Dec 03 00:12:56 crc kubenswrapper[4953]: I1203 00:12:56.206509 4953 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b33612b4-02d4-4f26-86bb-73a630d1ee58-client-ca\") on node \"crc\" DevicePath \"\"" Dec 03 00:12:56 crc kubenswrapper[4953]: I1203 00:12:56.206518 4953 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/90247294-af78-4d04-8aed-af90f93a034a-config\") on node \"crc\" DevicePath \"\"" Dec 03 00:12:56 crc kubenswrapper[4953]: I1203 00:12:56.206527 4953 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/90247294-af78-4d04-8aed-af90f93a034a-client-ca\") on node \"crc\" DevicePath \"\"" Dec 03 00:12:56 crc kubenswrapper[4953]: I1203 00:12:56.206535 4953 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fwqw4\" (UniqueName: \"kubernetes.io/projected/90247294-af78-4d04-8aed-af90f93a034a-kube-api-access-fwqw4\") on node \"crc\" DevicePath \"\"" Dec 03 00:12:56 crc kubenswrapper[4953]: I1203 00:12:56.206542 4953 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/90247294-af78-4d04-8aed-af90f93a034a-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 00:12:56 crc kubenswrapper[4953]: I1203 00:12:56.206550 4953 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b33612b4-02d4-4f26-86bb-73a630d1ee58-config\") on node \"crc\" DevicePath \"\"" Dec 03 00:12:56 crc kubenswrapper[4953]: I1203 00:12:56.559748 4953 generic.go:334] "Generic (PLEG): container finished" podID="b33612b4-02d4-4f26-86bb-73a630d1ee58" containerID="2027e38371dd6192fed82e47860af430e69b879668794026dc7f3b6ba304f8e0" exitCode=0 Dec 03 00:12:56 crc kubenswrapper[4953]: I1203 00:12:56.559845 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-l7sj9" event={"ID":"b33612b4-02d4-4f26-86bb-73a630d1ee58","Type":"ContainerDied","Data":"2027e38371dd6192fed82e47860af430e69b879668794026dc7f3b6ba304f8e0"} Dec 03 00:12:56 crc kubenswrapper[4953]: I1203 00:12:56.559867 4953 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-l7sj9" Dec 03 00:12:56 crc kubenswrapper[4953]: I1203 00:12:56.559897 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-l7sj9" event={"ID":"b33612b4-02d4-4f26-86bb-73a630d1ee58","Type":"ContainerDied","Data":"5a263737003b51158ff1c32ac37bbf8e444695110c8cf6d655006e58fc7c0938"} Dec 03 00:12:56 crc kubenswrapper[4953]: I1203 00:12:56.559926 4953 scope.go:117] "RemoveContainer" containerID="2027e38371dd6192fed82e47860af430e69b879668794026dc7f3b6ba304f8e0" Dec 03 00:12:56 crc kubenswrapper[4953]: I1203 00:12:56.563471 4953 generic.go:334] "Generic (PLEG): container finished" podID="90247294-af78-4d04-8aed-af90f93a034a" containerID="5146c8aac8d7e3cb2d34f3059cb8805fa3cfd735cbdd42ef3ea7e72e404d7cce" exitCode=0 Dec 03 00:12:56 crc kubenswrapper[4953]: I1203 00:12:56.563534 4953 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-52hrr" Dec 03 00:12:56 crc kubenswrapper[4953]: I1203 00:12:56.563543 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-52hrr" event={"ID":"90247294-af78-4d04-8aed-af90f93a034a","Type":"ContainerDied","Data":"5146c8aac8d7e3cb2d34f3059cb8805fa3cfd735cbdd42ef3ea7e72e404d7cce"} Dec 03 00:12:56 crc kubenswrapper[4953]: I1203 00:12:56.563591 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-52hrr" event={"ID":"90247294-af78-4d04-8aed-af90f93a034a","Type":"ContainerDied","Data":"7526cbddebe178ec2171c80e10fe7a7efc3d2628ee342a668d58be317ca48031"} Dec 03 00:12:56 crc kubenswrapper[4953]: I1203 00:12:56.592709 4953 scope.go:117] "RemoveContainer" containerID="2027e38371dd6192fed82e47860af430e69b879668794026dc7f3b6ba304f8e0" Dec 03 00:12:56 crc kubenswrapper[4953]: E1203 00:12:56.593346 4953 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2027e38371dd6192fed82e47860af430e69b879668794026dc7f3b6ba304f8e0\": container with ID starting with 2027e38371dd6192fed82e47860af430e69b879668794026dc7f3b6ba304f8e0 not found: ID does not exist" containerID="2027e38371dd6192fed82e47860af430e69b879668794026dc7f3b6ba304f8e0" Dec 03 00:12:56 crc kubenswrapper[4953]: I1203 00:12:56.593491 4953 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2027e38371dd6192fed82e47860af430e69b879668794026dc7f3b6ba304f8e0"} err="failed to get container status \"2027e38371dd6192fed82e47860af430e69b879668794026dc7f3b6ba304f8e0\": rpc error: code = NotFound desc = could not find container \"2027e38371dd6192fed82e47860af430e69b879668794026dc7f3b6ba304f8e0\": container with ID starting with 2027e38371dd6192fed82e47860af430e69b879668794026dc7f3b6ba304f8e0 not found: ID does not exist" Dec 03 00:12:56 crc kubenswrapper[4953]: I1203 00:12:56.593542 4953 scope.go:117] "RemoveContainer" containerID="5146c8aac8d7e3cb2d34f3059cb8805fa3cfd735cbdd42ef3ea7e72e404d7cce" Dec 03 00:12:56 crc kubenswrapper[4953]: I1203 00:12:56.615834 4953 scope.go:117] "RemoveContainer" containerID="5146c8aac8d7e3cb2d34f3059cb8805fa3cfd735cbdd42ef3ea7e72e404d7cce" Dec 03 00:12:56 crc kubenswrapper[4953]: E1203 00:12:56.616611 4953 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5146c8aac8d7e3cb2d34f3059cb8805fa3cfd735cbdd42ef3ea7e72e404d7cce\": container with ID starting with 5146c8aac8d7e3cb2d34f3059cb8805fa3cfd735cbdd42ef3ea7e72e404d7cce not found: ID does not exist" containerID="5146c8aac8d7e3cb2d34f3059cb8805fa3cfd735cbdd42ef3ea7e72e404d7cce" Dec 03 00:12:56 crc kubenswrapper[4953]: I1203 00:12:56.616642 4953 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5146c8aac8d7e3cb2d34f3059cb8805fa3cfd735cbdd42ef3ea7e72e404d7cce"} err="failed to get container status \"5146c8aac8d7e3cb2d34f3059cb8805fa3cfd735cbdd42ef3ea7e72e404d7cce\": rpc error: code = NotFound desc = could not find container \"5146c8aac8d7e3cb2d34f3059cb8805fa3cfd735cbdd42ef3ea7e72e404d7cce\": container with ID starting with 5146c8aac8d7e3cb2d34f3059cb8805fa3cfd735cbdd42ef3ea7e72e404d7cce not found: ID does not exist" Dec 03 00:12:56 crc kubenswrapper[4953]: I1203 00:12:56.625918 4953 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-52hrr"] Dec 03 00:12:56 crc kubenswrapper[4953]: I1203 00:12:56.629516 4953 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-52hrr"] Dec 03 00:12:56 crc kubenswrapper[4953]: I1203 00:12:56.646620 4953 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-l7sj9"] Dec 03 00:12:56 crc kubenswrapper[4953]: I1203 00:12:56.654160 4953 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-l7sj9"] Dec 03 00:12:57 crc kubenswrapper[4953]: I1203 00:12:57.130000 4953 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7447957dcb-m2cwp"] Dec 03 00:12:57 crc kubenswrapper[4953]: E1203 00:12:57.130384 4953 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="90247294-af78-4d04-8aed-af90f93a034a" containerName="route-controller-manager" Dec 03 00:12:57 crc kubenswrapper[4953]: I1203 00:12:57.130401 4953 state_mem.go:107] "Deleted CPUSet assignment" podUID="90247294-af78-4d04-8aed-af90f93a034a" containerName="route-controller-manager" Dec 03 00:12:57 crc kubenswrapper[4953]: E1203 00:12:57.130427 4953 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b33612b4-02d4-4f26-86bb-73a630d1ee58" containerName="controller-manager" Dec 03 00:12:57 crc kubenswrapper[4953]: I1203 00:12:57.130436 4953 state_mem.go:107] "Deleted CPUSet assignment" podUID="b33612b4-02d4-4f26-86bb-73a630d1ee58" containerName="controller-manager" Dec 03 00:12:57 crc kubenswrapper[4953]: I1203 00:12:57.130540 4953 memory_manager.go:354] "RemoveStaleState removing state" podUID="b33612b4-02d4-4f26-86bb-73a630d1ee58" containerName="controller-manager" Dec 03 00:12:57 crc kubenswrapper[4953]: I1203 00:12:57.130567 4953 memory_manager.go:354] "RemoveStaleState removing state" podUID="90247294-af78-4d04-8aed-af90f93a034a" containerName="route-controller-manager" Dec 03 00:12:57 crc kubenswrapper[4953]: I1203 00:12:57.131041 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-7447957dcb-m2cwp" Dec 03 00:12:57 crc kubenswrapper[4953]: I1203 00:12:57.134162 4953 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-58c9fbdd4b-5w4p8"] Dec 03 00:12:57 crc kubenswrapper[4953]: I1203 00:12:57.135390 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-58c9fbdd4b-5w4p8" Dec 03 00:12:57 crc kubenswrapper[4953]: I1203 00:12:57.135789 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Dec 03 00:12:57 crc kubenswrapper[4953]: I1203 00:12:57.136803 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Dec 03 00:12:57 crc kubenswrapper[4953]: I1203 00:12:57.137008 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Dec 03 00:12:57 crc kubenswrapper[4953]: I1203 00:12:57.137065 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Dec 03 00:12:57 crc kubenswrapper[4953]: I1203 00:12:57.137303 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Dec 03 00:12:57 crc kubenswrapper[4953]: I1203 00:12:57.138922 4953 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Dec 03 00:12:57 crc kubenswrapper[4953]: I1203 00:12:57.139294 4953 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Dec 03 00:12:57 crc kubenswrapper[4953]: I1203 00:12:57.140300 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Dec 03 00:12:57 crc kubenswrapper[4953]: I1203 00:12:57.140557 4953 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Dec 03 00:12:57 crc kubenswrapper[4953]: I1203 00:12:57.142528 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Dec 03 00:12:57 crc kubenswrapper[4953]: I1203 00:12:57.144380 4953 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Dec 03 00:12:57 crc kubenswrapper[4953]: I1203 00:12:57.145857 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Dec 03 00:12:57 crc kubenswrapper[4953]: I1203 00:12:57.146059 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Dec 03 00:12:57 crc kubenswrapper[4953]: I1203 00:12:57.148872 4953 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7447957dcb-m2cwp"] Dec 03 00:12:57 crc kubenswrapper[4953]: I1203 00:12:57.153116 4953 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-58c9fbdd4b-5w4p8"] Dec 03 00:12:57 crc kubenswrapper[4953]: I1203 00:12:57.248304 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fe523418-5f1d-472a-bf22-246672d50480-config\") pod \"route-controller-manager-7447957dcb-m2cwp\" (UID: \"fe523418-5f1d-472a-bf22-246672d50480\") " pod="openshift-route-controller-manager/route-controller-manager-7447957dcb-m2cwp" Dec 03 00:12:57 crc kubenswrapper[4953]: I1203 00:12:57.248865 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/db3140f8-1095-46a1-a8e5-e6421ca0dfb3-proxy-ca-bundles\") pod \"controller-manager-58c9fbdd4b-5w4p8\" (UID: \"db3140f8-1095-46a1-a8e5-e6421ca0dfb3\") " pod="openshift-controller-manager/controller-manager-58c9fbdd4b-5w4p8" Dec 03 00:12:57 crc kubenswrapper[4953]: I1203 00:12:57.248906 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gcgd4\" (UniqueName: \"kubernetes.io/projected/db3140f8-1095-46a1-a8e5-e6421ca0dfb3-kube-api-access-gcgd4\") pod \"controller-manager-58c9fbdd4b-5w4p8\" (UID: \"db3140f8-1095-46a1-a8e5-e6421ca0dfb3\") " pod="openshift-controller-manager/controller-manager-58c9fbdd4b-5w4p8" Dec 03 00:12:57 crc kubenswrapper[4953]: I1203 00:12:57.249002 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fe523418-5f1d-472a-bf22-246672d50480-serving-cert\") pod \"route-controller-manager-7447957dcb-m2cwp\" (UID: \"fe523418-5f1d-472a-bf22-246672d50480\") " pod="openshift-route-controller-manager/route-controller-manager-7447957dcb-m2cwp" Dec 03 00:12:57 crc kubenswrapper[4953]: I1203 00:12:57.249050 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/db3140f8-1095-46a1-a8e5-e6421ca0dfb3-config\") pod \"controller-manager-58c9fbdd4b-5w4p8\" (UID: \"db3140f8-1095-46a1-a8e5-e6421ca0dfb3\") " pod="openshift-controller-manager/controller-manager-58c9fbdd4b-5w4p8" Dec 03 00:12:57 crc kubenswrapper[4953]: I1203 00:12:57.249084 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/db3140f8-1095-46a1-a8e5-e6421ca0dfb3-client-ca\") pod \"controller-manager-58c9fbdd4b-5w4p8\" (UID: \"db3140f8-1095-46a1-a8e5-e6421ca0dfb3\") " pod="openshift-controller-manager/controller-manager-58c9fbdd4b-5w4p8" Dec 03 00:12:57 crc kubenswrapper[4953]: I1203 00:12:57.249110 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/db3140f8-1095-46a1-a8e5-e6421ca0dfb3-serving-cert\") pod \"controller-manager-58c9fbdd4b-5w4p8\" (UID: \"db3140f8-1095-46a1-a8e5-e6421ca0dfb3\") " pod="openshift-controller-manager/controller-manager-58c9fbdd4b-5w4p8" Dec 03 00:12:57 crc kubenswrapper[4953]: I1203 00:12:57.249138 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/fe523418-5f1d-472a-bf22-246672d50480-client-ca\") pod \"route-controller-manager-7447957dcb-m2cwp\" (UID: \"fe523418-5f1d-472a-bf22-246672d50480\") " pod="openshift-route-controller-manager/route-controller-manager-7447957dcb-m2cwp" Dec 03 00:12:57 crc kubenswrapper[4953]: I1203 00:12:57.249312 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qfrlj\" (UniqueName: \"kubernetes.io/projected/fe523418-5f1d-472a-bf22-246672d50480-kube-api-access-qfrlj\") pod \"route-controller-manager-7447957dcb-m2cwp\" (UID: \"fe523418-5f1d-472a-bf22-246672d50480\") " pod="openshift-route-controller-manager/route-controller-manager-7447957dcb-m2cwp" Dec 03 00:12:57 crc kubenswrapper[4953]: I1203 00:12:57.350174 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fe523418-5f1d-472a-bf22-246672d50480-config\") pod \"route-controller-manager-7447957dcb-m2cwp\" (UID: \"fe523418-5f1d-472a-bf22-246672d50480\") " pod="openshift-route-controller-manager/route-controller-manager-7447957dcb-m2cwp" Dec 03 00:12:57 crc kubenswrapper[4953]: I1203 00:12:57.350232 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/db3140f8-1095-46a1-a8e5-e6421ca0dfb3-proxy-ca-bundles\") pod \"controller-manager-58c9fbdd4b-5w4p8\" (UID: \"db3140f8-1095-46a1-a8e5-e6421ca0dfb3\") " pod="openshift-controller-manager/controller-manager-58c9fbdd4b-5w4p8" Dec 03 00:12:57 crc kubenswrapper[4953]: I1203 00:12:57.350266 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gcgd4\" (UniqueName: \"kubernetes.io/projected/db3140f8-1095-46a1-a8e5-e6421ca0dfb3-kube-api-access-gcgd4\") pod \"controller-manager-58c9fbdd4b-5w4p8\" (UID: \"db3140f8-1095-46a1-a8e5-e6421ca0dfb3\") " pod="openshift-controller-manager/controller-manager-58c9fbdd4b-5w4p8" Dec 03 00:12:57 crc kubenswrapper[4953]: I1203 00:12:57.350329 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fe523418-5f1d-472a-bf22-246672d50480-serving-cert\") pod \"route-controller-manager-7447957dcb-m2cwp\" (UID: \"fe523418-5f1d-472a-bf22-246672d50480\") " pod="openshift-route-controller-manager/route-controller-manager-7447957dcb-m2cwp" Dec 03 00:12:57 crc kubenswrapper[4953]: I1203 00:12:57.350376 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/db3140f8-1095-46a1-a8e5-e6421ca0dfb3-config\") pod \"controller-manager-58c9fbdd4b-5w4p8\" (UID: \"db3140f8-1095-46a1-a8e5-e6421ca0dfb3\") " pod="openshift-controller-manager/controller-manager-58c9fbdd4b-5w4p8" Dec 03 00:12:57 crc kubenswrapper[4953]: I1203 00:12:57.350397 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/db3140f8-1095-46a1-a8e5-e6421ca0dfb3-client-ca\") pod \"controller-manager-58c9fbdd4b-5w4p8\" (UID: \"db3140f8-1095-46a1-a8e5-e6421ca0dfb3\") " pod="openshift-controller-manager/controller-manager-58c9fbdd4b-5w4p8" Dec 03 00:12:57 crc kubenswrapper[4953]: I1203 00:12:57.350415 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/db3140f8-1095-46a1-a8e5-e6421ca0dfb3-serving-cert\") pod \"controller-manager-58c9fbdd4b-5w4p8\" (UID: \"db3140f8-1095-46a1-a8e5-e6421ca0dfb3\") " pod="openshift-controller-manager/controller-manager-58c9fbdd4b-5w4p8" Dec 03 00:12:57 crc kubenswrapper[4953]: I1203 00:12:57.350434 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/fe523418-5f1d-472a-bf22-246672d50480-client-ca\") pod \"route-controller-manager-7447957dcb-m2cwp\" (UID: \"fe523418-5f1d-472a-bf22-246672d50480\") " pod="openshift-route-controller-manager/route-controller-manager-7447957dcb-m2cwp" Dec 03 00:12:57 crc kubenswrapper[4953]: I1203 00:12:57.350480 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qfrlj\" (UniqueName: \"kubernetes.io/projected/fe523418-5f1d-472a-bf22-246672d50480-kube-api-access-qfrlj\") pod \"route-controller-manager-7447957dcb-m2cwp\" (UID: \"fe523418-5f1d-472a-bf22-246672d50480\") " pod="openshift-route-controller-manager/route-controller-manager-7447957dcb-m2cwp" Dec 03 00:12:57 crc kubenswrapper[4953]: I1203 00:12:57.351658 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fe523418-5f1d-472a-bf22-246672d50480-config\") pod \"route-controller-manager-7447957dcb-m2cwp\" (UID: \"fe523418-5f1d-472a-bf22-246672d50480\") " pod="openshift-route-controller-manager/route-controller-manager-7447957dcb-m2cwp" Dec 03 00:12:57 crc kubenswrapper[4953]: I1203 00:12:57.351823 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/fe523418-5f1d-472a-bf22-246672d50480-client-ca\") pod \"route-controller-manager-7447957dcb-m2cwp\" (UID: \"fe523418-5f1d-472a-bf22-246672d50480\") " pod="openshift-route-controller-manager/route-controller-manager-7447957dcb-m2cwp" Dec 03 00:12:57 crc kubenswrapper[4953]: I1203 00:12:57.352257 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/db3140f8-1095-46a1-a8e5-e6421ca0dfb3-config\") pod \"controller-manager-58c9fbdd4b-5w4p8\" (UID: \"db3140f8-1095-46a1-a8e5-e6421ca0dfb3\") " pod="openshift-controller-manager/controller-manager-58c9fbdd4b-5w4p8" Dec 03 00:12:57 crc kubenswrapper[4953]: I1203 00:12:57.352591 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/db3140f8-1095-46a1-a8e5-e6421ca0dfb3-proxy-ca-bundles\") pod \"controller-manager-58c9fbdd4b-5w4p8\" (UID: \"db3140f8-1095-46a1-a8e5-e6421ca0dfb3\") " pod="openshift-controller-manager/controller-manager-58c9fbdd4b-5w4p8" Dec 03 00:12:57 crc kubenswrapper[4953]: I1203 00:12:57.353752 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/db3140f8-1095-46a1-a8e5-e6421ca0dfb3-client-ca\") pod \"controller-manager-58c9fbdd4b-5w4p8\" (UID: \"db3140f8-1095-46a1-a8e5-e6421ca0dfb3\") " pod="openshift-controller-manager/controller-manager-58c9fbdd4b-5w4p8" Dec 03 00:12:57 crc kubenswrapper[4953]: I1203 00:12:57.358237 4953 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="90247294-af78-4d04-8aed-af90f93a034a" path="/var/lib/kubelet/pods/90247294-af78-4d04-8aed-af90f93a034a/volumes" Dec 03 00:12:57 crc kubenswrapper[4953]: I1203 00:12:57.358482 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fe523418-5f1d-472a-bf22-246672d50480-serving-cert\") pod \"route-controller-manager-7447957dcb-m2cwp\" (UID: \"fe523418-5f1d-472a-bf22-246672d50480\") " pod="openshift-route-controller-manager/route-controller-manager-7447957dcb-m2cwp" Dec 03 00:12:57 crc kubenswrapper[4953]: I1203 00:12:57.358494 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/db3140f8-1095-46a1-a8e5-e6421ca0dfb3-serving-cert\") pod \"controller-manager-58c9fbdd4b-5w4p8\" (UID: \"db3140f8-1095-46a1-a8e5-e6421ca0dfb3\") " pod="openshift-controller-manager/controller-manager-58c9fbdd4b-5w4p8" Dec 03 00:12:57 crc kubenswrapper[4953]: I1203 00:12:57.359132 4953 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b33612b4-02d4-4f26-86bb-73a630d1ee58" path="/var/lib/kubelet/pods/b33612b4-02d4-4f26-86bb-73a630d1ee58/volumes" Dec 03 00:12:57 crc kubenswrapper[4953]: I1203 00:12:57.370623 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qfrlj\" (UniqueName: \"kubernetes.io/projected/fe523418-5f1d-472a-bf22-246672d50480-kube-api-access-qfrlj\") pod \"route-controller-manager-7447957dcb-m2cwp\" (UID: \"fe523418-5f1d-472a-bf22-246672d50480\") " pod="openshift-route-controller-manager/route-controller-manager-7447957dcb-m2cwp" Dec 03 00:12:57 crc kubenswrapper[4953]: I1203 00:12:57.372588 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gcgd4\" (UniqueName: \"kubernetes.io/projected/db3140f8-1095-46a1-a8e5-e6421ca0dfb3-kube-api-access-gcgd4\") pod \"controller-manager-58c9fbdd4b-5w4p8\" (UID: \"db3140f8-1095-46a1-a8e5-e6421ca0dfb3\") " pod="openshift-controller-manager/controller-manager-58c9fbdd4b-5w4p8" Dec 03 00:12:57 crc kubenswrapper[4953]: I1203 00:12:57.460799 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-7447957dcb-m2cwp" Dec 03 00:12:57 crc kubenswrapper[4953]: I1203 00:12:57.475134 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-58c9fbdd4b-5w4p8" Dec 03 00:12:57 crc kubenswrapper[4953]: I1203 00:12:57.712387 4953 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-58c9fbdd4b-5w4p8"] Dec 03 00:12:57 crc kubenswrapper[4953]: I1203 00:12:57.879186 4953 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7447957dcb-m2cwp"] Dec 03 00:12:57 crc kubenswrapper[4953]: W1203 00:12:57.881011 4953 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podfe523418_5f1d_472a_bf22_246672d50480.slice/crio-796b1530b17e4b7ecf299cb7af0492a118cafda3857b9b32eb3a73056244e8c9 WatchSource:0}: Error finding container 796b1530b17e4b7ecf299cb7af0492a118cafda3857b9b32eb3a73056244e8c9: Status 404 returned error can't find the container with id 796b1530b17e4b7ecf299cb7af0492a118cafda3857b9b32eb3a73056244e8c9 Dec 03 00:12:58 crc kubenswrapper[4953]: I1203 00:12:58.612346 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-7447957dcb-m2cwp" event={"ID":"fe523418-5f1d-472a-bf22-246672d50480","Type":"ContainerStarted","Data":"87d1955f38cf491a4e593c5b87b7335bc3ed7e5bbe7ddfe055fe99731e169aac"} Dec 03 00:12:58 crc kubenswrapper[4953]: I1203 00:12:58.612965 4953 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-7447957dcb-m2cwp" Dec 03 00:12:58 crc kubenswrapper[4953]: I1203 00:12:58.612998 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-7447957dcb-m2cwp" event={"ID":"fe523418-5f1d-472a-bf22-246672d50480","Type":"ContainerStarted","Data":"796b1530b17e4b7ecf299cb7af0492a118cafda3857b9b32eb3a73056244e8c9"} Dec 03 00:12:58 crc kubenswrapper[4953]: I1203 00:12:58.616377 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-58c9fbdd4b-5w4p8" event={"ID":"db3140f8-1095-46a1-a8e5-e6421ca0dfb3","Type":"ContainerStarted","Data":"5112cf9978066d88161e4e65533e19cb4dbbc230aec4f252851be66ab554428b"} Dec 03 00:12:58 crc kubenswrapper[4953]: I1203 00:12:58.616422 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-58c9fbdd4b-5w4p8" event={"ID":"db3140f8-1095-46a1-a8e5-e6421ca0dfb3","Type":"ContainerStarted","Data":"ce3d8c4a0dbfae3c1b65ed1ce3ff4b956b5383594ece83addd261c3705a5df27"} Dec 03 00:12:58 crc kubenswrapper[4953]: I1203 00:12:58.616673 4953 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-58c9fbdd4b-5w4p8" Dec 03 00:12:58 crc kubenswrapper[4953]: I1203 00:12:58.619535 4953 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-7447957dcb-m2cwp" Dec 03 00:12:58 crc kubenswrapper[4953]: I1203 00:12:58.622906 4953 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-58c9fbdd4b-5w4p8" Dec 03 00:12:58 crc kubenswrapper[4953]: I1203 00:12:58.642872 4953 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-7447957dcb-m2cwp" podStartSLOduration=2.642842094 podStartE2EDuration="2.642842094s" podCreationTimestamp="2025-12-03 00:12:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 00:12:58.636940406 +0000 UTC m=+367.621221199" watchObservedRunningTime="2025-12-03 00:12:58.642842094 +0000 UTC m=+367.627122877" Dec 03 00:12:58 crc kubenswrapper[4953]: I1203 00:12:58.692786 4953 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-58c9fbdd4b-5w4p8" podStartSLOduration=3.692755361 podStartE2EDuration="3.692755361s" podCreationTimestamp="2025-12-03 00:12:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 00:12:58.667670534 +0000 UTC m=+367.651951347" watchObservedRunningTime="2025-12-03 00:12:58.692755361 +0000 UTC m=+367.677036144" Dec 03 00:13:18 crc kubenswrapper[4953]: I1203 00:13:18.945169 4953 patch_prober.go:28] interesting pod/machine-config-daemon-6dfbn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 00:13:18 crc kubenswrapper[4953]: I1203 00:13:18.946176 4953 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-6dfbn" podUID="53ffb1e9-9a2e-4869-80d7-776f4aa7ef8b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 00:13:35 crc kubenswrapper[4953]: I1203 00:13:35.520016 4953 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-58c9fbdd4b-5w4p8"] Dec 03 00:13:35 crc kubenswrapper[4953]: I1203 00:13:35.521274 4953 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-58c9fbdd4b-5w4p8" podUID="db3140f8-1095-46a1-a8e5-e6421ca0dfb3" containerName="controller-manager" containerID="cri-o://5112cf9978066d88161e4e65533e19cb4dbbc230aec4f252851be66ab554428b" gracePeriod=30 Dec 03 00:13:35 crc kubenswrapper[4953]: I1203 00:13:35.860177 4953 generic.go:334] "Generic (PLEG): container finished" podID="db3140f8-1095-46a1-a8e5-e6421ca0dfb3" containerID="5112cf9978066d88161e4e65533e19cb4dbbc230aec4f252851be66ab554428b" exitCode=0 Dec 03 00:13:35 crc kubenswrapper[4953]: I1203 00:13:35.860251 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-58c9fbdd4b-5w4p8" event={"ID":"db3140f8-1095-46a1-a8e5-e6421ca0dfb3","Type":"ContainerDied","Data":"5112cf9978066d88161e4e65533e19cb4dbbc230aec4f252851be66ab554428b"} Dec 03 00:13:36 crc kubenswrapper[4953]: I1203 00:13:36.021369 4953 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-58c9fbdd4b-5w4p8" Dec 03 00:13:36 crc kubenswrapper[4953]: I1203 00:13:36.122346 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gcgd4\" (UniqueName: \"kubernetes.io/projected/db3140f8-1095-46a1-a8e5-e6421ca0dfb3-kube-api-access-gcgd4\") pod \"db3140f8-1095-46a1-a8e5-e6421ca0dfb3\" (UID: \"db3140f8-1095-46a1-a8e5-e6421ca0dfb3\") " Dec 03 00:13:36 crc kubenswrapper[4953]: I1203 00:13:36.122525 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/db3140f8-1095-46a1-a8e5-e6421ca0dfb3-config\") pod \"db3140f8-1095-46a1-a8e5-e6421ca0dfb3\" (UID: \"db3140f8-1095-46a1-a8e5-e6421ca0dfb3\") " Dec 03 00:13:36 crc kubenswrapper[4953]: I1203 00:13:36.122555 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/db3140f8-1095-46a1-a8e5-e6421ca0dfb3-serving-cert\") pod \"db3140f8-1095-46a1-a8e5-e6421ca0dfb3\" (UID: \"db3140f8-1095-46a1-a8e5-e6421ca0dfb3\") " Dec 03 00:13:36 crc kubenswrapper[4953]: I1203 00:13:36.122656 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/db3140f8-1095-46a1-a8e5-e6421ca0dfb3-proxy-ca-bundles\") pod \"db3140f8-1095-46a1-a8e5-e6421ca0dfb3\" (UID: \"db3140f8-1095-46a1-a8e5-e6421ca0dfb3\") " Dec 03 00:13:36 crc kubenswrapper[4953]: I1203 00:13:36.122760 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/db3140f8-1095-46a1-a8e5-e6421ca0dfb3-client-ca\") pod \"db3140f8-1095-46a1-a8e5-e6421ca0dfb3\" (UID: \"db3140f8-1095-46a1-a8e5-e6421ca0dfb3\") " Dec 03 00:13:36 crc kubenswrapper[4953]: I1203 00:13:36.123811 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/db3140f8-1095-46a1-a8e5-e6421ca0dfb3-client-ca" (OuterVolumeSpecName: "client-ca") pod "db3140f8-1095-46a1-a8e5-e6421ca0dfb3" (UID: "db3140f8-1095-46a1-a8e5-e6421ca0dfb3"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:13:36 crc kubenswrapper[4953]: I1203 00:13:36.124300 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/db3140f8-1095-46a1-a8e5-e6421ca0dfb3-config" (OuterVolumeSpecName: "config") pod "db3140f8-1095-46a1-a8e5-e6421ca0dfb3" (UID: "db3140f8-1095-46a1-a8e5-e6421ca0dfb3"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:13:36 crc kubenswrapper[4953]: I1203 00:13:36.124500 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/db3140f8-1095-46a1-a8e5-e6421ca0dfb3-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "db3140f8-1095-46a1-a8e5-e6421ca0dfb3" (UID: "db3140f8-1095-46a1-a8e5-e6421ca0dfb3"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:13:36 crc kubenswrapper[4953]: I1203 00:13:36.129601 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/db3140f8-1095-46a1-a8e5-e6421ca0dfb3-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "db3140f8-1095-46a1-a8e5-e6421ca0dfb3" (UID: "db3140f8-1095-46a1-a8e5-e6421ca0dfb3"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:13:36 crc kubenswrapper[4953]: I1203 00:13:36.129671 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/db3140f8-1095-46a1-a8e5-e6421ca0dfb3-kube-api-access-gcgd4" (OuterVolumeSpecName: "kube-api-access-gcgd4") pod "db3140f8-1095-46a1-a8e5-e6421ca0dfb3" (UID: "db3140f8-1095-46a1-a8e5-e6421ca0dfb3"). InnerVolumeSpecName "kube-api-access-gcgd4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:13:36 crc kubenswrapper[4953]: I1203 00:13:36.224682 4953 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/db3140f8-1095-46a1-a8e5-e6421ca0dfb3-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Dec 03 00:13:36 crc kubenswrapper[4953]: I1203 00:13:36.224732 4953 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/db3140f8-1095-46a1-a8e5-e6421ca0dfb3-client-ca\") on node \"crc\" DevicePath \"\"" Dec 03 00:13:36 crc kubenswrapper[4953]: I1203 00:13:36.224747 4953 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gcgd4\" (UniqueName: \"kubernetes.io/projected/db3140f8-1095-46a1-a8e5-e6421ca0dfb3-kube-api-access-gcgd4\") on node \"crc\" DevicePath \"\"" Dec 03 00:13:36 crc kubenswrapper[4953]: I1203 00:13:36.224764 4953 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/db3140f8-1095-46a1-a8e5-e6421ca0dfb3-config\") on node \"crc\" DevicePath \"\"" Dec 03 00:13:36 crc kubenswrapper[4953]: I1203 00:13:36.224774 4953 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/db3140f8-1095-46a1-a8e5-e6421ca0dfb3-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 00:13:36 crc kubenswrapper[4953]: I1203 00:13:36.868276 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-58c9fbdd4b-5w4p8" event={"ID":"db3140f8-1095-46a1-a8e5-e6421ca0dfb3","Type":"ContainerDied","Data":"ce3d8c4a0dbfae3c1b65ed1ce3ff4b956b5383594ece83addd261c3705a5df27"} Dec 03 00:13:36 crc kubenswrapper[4953]: I1203 00:13:36.868341 4953 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-58c9fbdd4b-5w4p8" Dec 03 00:13:36 crc kubenswrapper[4953]: I1203 00:13:36.868360 4953 scope.go:117] "RemoveContainer" containerID="5112cf9978066d88161e4e65533e19cb4dbbc230aec4f252851be66ab554428b" Dec 03 00:13:36 crc kubenswrapper[4953]: I1203 00:13:36.923377 4953 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-58c9fbdd4b-5w4p8"] Dec 03 00:13:36 crc kubenswrapper[4953]: I1203 00:13:36.932374 4953 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-58c9fbdd4b-5w4p8"] Dec 03 00:13:37 crc kubenswrapper[4953]: I1203 00:13:37.155307 4953 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-556b846d59-5cf9d"] Dec 03 00:13:37 crc kubenswrapper[4953]: E1203 00:13:37.156038 4953 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="db3140f8-1095-46a1-a8e5-e6421ca0dfb3" containerName="controller-manager" Dec 03 00:13:37 crc kubenswrapper[4953]: I1203 00:13:37.156140 4953 state_mem.go:107] "Deleted CPUSet assignment" podUID="db3140f8-1095-46a1-a8e5-e6421ca0dfb3" containerName="controller-manager" Dec 03 00:13:37 crc kubenswrapper[4953]: I1203 00:13:37.156335 4953 memory_manager.go:354] "RemoveStaleState removing state" podUID="db3140f8-1095-46a1-a8e5-e6421ca0dfb3" containerName="controller-manager" Dec 03 00:13:37 crc kubenswrapper[4953]: I1203 00:13:37.156933 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-556b846d59-5cf9d" Dec 03 00:13:37 crc kubenswrapper[4953]: I1203 00:13:37.159140 4953 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Dec 03 00:13:37 crc kubenswrapper[4953]: I1203 00:13:37.159255 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Dec 03 00:13:37 crc kubenswrapper[4953]: I1203 00:13:37.159512 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Dec 03 00:13:37 crc kubenswrapper[4953]: I1203 00:13:37.160257 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Dec 03 00:13:37 crc kubenswrapper[4953]: I1203 00:13:37.162432 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Dec 03 00:13:37 crc kubenswrapper[4953]: I1203 00:13:37.162830 4953 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Dec 03 00:13:37 crc kubenswrapper[4953]: I1203 00:13:37.165778 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Dec 03 00:13:37 crc kubenswrapper[4953]: I1203 00:13:37.176814 4953 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-556b846d59-5cf9d"] Dec 03 00:13:37 crc kubenswrapper[4953]: I1203 00:13:37.240074 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f677c67f-7084-4e3a-9255-404b4dcd8311-serving-cert\") pod \"controller-manager-556b846d59-5cf9d\" (UID: \"f677c67f-7084-4e3a-9255-404b4dcd8311\") " pod="openshift-controller-manager/controller-manager-556b846d59-5cf9d" Dec 03 00:13:37 crc kubenswrapper[4953]: I1203 00:13:37.240628 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f677c67f-7084-4e3a-9255-404b4dcd8311-config\") pod \"controller-manager-556b846d59-5cf9d\" (UID: \"f677c67f-7084-4e3a-9255-404b4dcd8311\") " pod="openshift-controller-manager/controller-manager-556b846d59-5cf9d" Dec 03 00:13:37 crc kubenswrapper[4953]: I1203 00:13:37.240816 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/f677c67f-7084-4e3a-9255-404b4dcd8311-proxy-ca-bundles\") pod \"controller-manager-556b846d59-5cf9d\" (UID: \"f677c67f-7084-4e3a-9255-404b4dcd8311\") " pod="openshift-controller-manager/controller-manager-556b846d59-5cf9d" Dec 03 00:13:37 crc kubenswrapper[4953]: I1203 00:13:37.241072 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/f677c67f-7084-4e3a-9255-404b4dcd8311-client-ca\") pod \"controller-manager-556b846d59-5cf9d\" (UID: \"f677c67f-7084-4e3a-9255-404b4dcd8311\") " pod="openshift-controller-manager/controller-manager-556b846d59-5cf9d" Dec 03 00:13:37 crc kubenswrapper[4953]: I1203 00:13:37.241284 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pr55q\" (UniqueName: \"kubernetes.io/projected/f677c67f-7084-4e3a-9255-404b4dcd8311-kube-api-access-pr55q\") pod \"controller-manager-556b846d59-5cf9d\" (UID: \"f677c67f-7084-4e3a-9255-404b4dcd8311\") " pod="openshift-controller-manager/controller-manager-556b846d59-5cf9d" Dec 03 00:13:37 crc kubenswrapper[4953]: I1203 00:13:37.344419 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/f677c67f-7084-4e3a-9255-404b4dcd8311-proxy-ca-bundles\") pod \"controller-manager-556b846d59-5cf9d\" (UID: \"f677c67f-7084-4e3a-9255-404b4dcd8311\") " pod="openshift-controller-manager/controller-manager-556b846d59-5cf9d" Dec 03 00:13:37 crc kubenswrapper[4953]: I1203 00:13:37.345080 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/f677c67f-7084-4e3a-9255-404b4dcd8311-client-ca\") pod \"controller-manager-556b846d59-5cf9d\" (UID: \"f677c67f-7084-4e3a-9255-404b4dcd8311\") " pod="openshift-controller-manager/controller-manager-556b846d59-5cf9d" Dec 03 00:13:37 crc kubenswrapper[4953]: I1203 00:13:37.346208 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pr55q\" (UniqueName: \"kubernetes.io/projected/f677c67f-7084-4e3a-9255-404b4dcd8311-kube-api-access-pr55q\") pod \"controller-manager-556b846d59-5cf9d\" (UID: \"f677c67f-7084-4e3a-9255-404b4dcd8311\") " pod="openshift-controller-manager/controller-manager-556b846d59-5cf9d" Dec 03 00:13:37 crc kubenswrapper[4953]: I1203 00:13:37.346305 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f677c67f-7084-4e3a-9255-404b4dcd8311-serving-cert\") pod \"controller-manager-556b846d59-5cf9d\" (UID: \"f677c67f-7084-4e3a-9255-404b4dcd8311\") " pod="openshift-controller-manager/controller-manager-556b846d59-5cf9d" Dec 03 00:13:37 crc kubenswrapper[4953]: I1203 00:13:37.346332 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f677c67f-7084-4e3a-9255-404b4dcd8311-config\") pod \"controller-manager-556b846d59-5cf9d\" (UID: \"f677c67f-7084-4e3a-9255-404b4dcd8311\") " pod="openshift-controller-manager/controller-manager-556b846d59-5cf9d" Dec 03 00:13:37 crc kubenswrapper[4953]: I1203 00:13:37.346217 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/f677c67f-7084-4e3a-9255-404b4dcd8311-proxy-ca-bundles\") pod \"controller-manager-556b846d59-5cf9d\" (UID: \"f677c67f-7084-4e3a-9255-404b4dcd8311\") " pod="openshift-controller-manager/controller-manager-556b846d59-5cf9d" Dec 03 00:13:37 crc kubenswrapper[4953]: I1203 00:13:37.346728 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/f677c67f-7084-4e3a-9255-404b4dcd8311-client-ca\") pod \"controller-manager-556b846d59-5cf9d\" (UID: \"f677c67f-7084-4e3a-9255-404b4dcd8311\") " pod="openshift-controller-manager/controller-manager-556b846d59-5cf9d" Dec 03 00:13:37 crc kubenswrapper[4953]: I1203 00:13:37.347634 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f677c67f-7084-4e3a-9255-404b4dcd8311-config\") pod \"controller-manager-556b846d59-5cf9d\" (UID: \"f677c67f-7084-4e3a-9255-404b4dcd8311\") " pod="openshift-controller-manager/controller-manager-556b846d59-5cf9d" Dec 03 00:13:37 crc kubenswrapper[4953]: I1203 00:13:37.352305 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f677c67f-7084-4e3a-9255-404b4dcd8311-serving-cert\") pod \"controller-manager-556b846d59-5cf9d\" (UID: \"f677c67f-7084-4e3a-9255-404b4dcd8311\") " pod="openshift-controller-manager/controller-manager-556b846d59-5cf9d" Dec 03 00:13:37 crc kubenswrapper[4953]: I1203 00:13:37.365120 4953 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="db3140f8-1095-46a1-a8e5-e6421ca0dfb3" path="/var/lib/kubelet/pods/db3140f8-1095-46a1-a8e5-e6421ca0dfb3/volumes" Dec 03 00:13:37 crc kubenswrapper[4953]: I1203 00:13:37.370500 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pr55q\" (UniqueName: \"kubernetes.io/projected/f677c67f-7084-4e3a-9255-404b4dcd8311-kube-api-access-pr55q\") pod \"controller-manager-556b846d59-5cf9d\" (UID: \"f677c67f-7084-4e3a-9255-404b4dcd8311\") " pod="openshift-controller-manager/controller-manager-556b846d59-5cf9d" Dec 03 00:13:37 crc kubenswrapper[4953]: I1203 00:13:37.514915 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-556b846d59-5cf9d" Dec 03 00:13:37 crc kubenswrapper[4953]: I1203 00:13:37.619098 4953 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-z74gr"] Dec 03 00:13:37 crc kubenswrapper[4953]: I1203 00:13:37.620219 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-z74gr" Dec 03 00:13:37 crc kubenswrapper[4953]: I1203 00:13:37.639500 4953 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-z74gr"] Dec 03 00:13:37 crc kubenswrapper[4953]: I1203 00:13:37.755193 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/7d8475bc-7ad5-4381-b5dd-b45a6b9f73e2-bound-sa-token\") pod \"image-registry-66df7c8f76-z74gr\" (UID: \"7d8475bc-7ad5-4381-b5dd-b45a6b9f73e2\") " pod="openshift-image-registry/image-registry-66df7c8f76-z74gr" Dec 03 00:13:37 crc kubenswrapper[4953]: I1203 00:13:37.755729 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/7d8475bc-7ad5-4381-b5dd-b45a6b9f73e2-installation-pull-secrets\") pod \"image-registry-66df7c8f76-z74gr\" (UID: \"7d8475bc-7ad5-4381-b5dd-b45a6b9f73e2\") " pod="openshift-image-registry/image-registry-66df7c8f76-z74gr" Dec 03 00:13:37 crc kubenswrapper[4953]: I1203 00:13:37.755770 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/7d8475bc-7ad5-4381-b5dd-b45a6b9f73e2-ca-trust-extracted\") pod \"image-registry-66df7c8f76-z74gr\" (UID: \"7d8475bc-7ad5-4381-b5dd-b45a6b9f73e2\") " pod="openshift-image-registry/image-registry-66df7c8f76-z74gr" Dec 03 00:13:37 crc kubenswrapper[4953]: I1203 00:13:37.755798 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/7d8475bc-7ad5-4381-b5dd-b45a6b9f73e2-trusted-ca\") pod \"image-registry-66df7c8f76-z74gr\" (UID: \"7d8475bc-7ad5-4381-b5dd-b45a6b9f73e2\") " pod="openshift-image-registry/image-registry-66df7c8f76-z74gr" Dec 03 00:13:37 crc kubenswrapper[4953]: I1203 00:13:37.755867 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/7d8475bc-7ad5-4381-b5dd-b45a6b9f73e2-registry-tls\") pod \"image-registry-66df7c8f76-z74gr\" (UID: \"7d8475bc-7ad5-4381-b5dd-b45a6b9f73e2\") " pod="openshift-image-registry/image-registry-66df7c8f76-z74gr" Dec 03 00:13:37 crc kubenswrapper[4953]: I1203 00:13:37.755896 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/7d8475bc-7ad5-4381-b5dd-b45a6b9f73e2-registry-certificates\") pod \"image-registry-66df7c8f76-z74gr\" (UID: \"7d8475bc-7ad5-4381-b5dd-b45a6b9f73e2\") " pod="openshift-image-registry/image-registry-66df7c8f76-z74gr" Dec 03 00:13:37 crc kubenswrapper[4953]: I1203 00:13:37.755914 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ct867\" (UniqueName: \"kubernetes.io/projected/7d8475bc-7ad5-4381-b5dd-b45a6b9f73e2-kube-api-access-ct867\") pod \"image-registry-66df7c8f76-z74gr\" (UID: \"7d8475bc-7ad5-4381-b5dd-b45a6b9f73e2\") " pod="openshift-image-registry/image-registry-66df7c8f76-z74gr" Dec 03 00:13:37 crc kubenswrapper[4953]: I1203 00:13:37.755944 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-z74gr\" (UID: \"7d8475bc-7ad5-4381-b5dd-b45a6b9f73e2\") " pod="openshift-image-registry/image-registry-66df7c8f76-z74gr" Dec 03 00:13:37 crc kubenswrapper[4953]: I1203 00:13:37.803599 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-z74gr\" (UID: \"7d8475bc-7ad5-4381-b5dd-b45a6b9f73e2\") " pod="openshift-image-registry/image-registry-66df7c8f76-z74gr" Dec 03 00:13:37 crc kubenswrapper[4953]: I1203 00:13:37.843690 4953 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-556b846d59-5cf9d"] Dec 03 00:13:37 crc kubenswrapper[4953]: I1203 00:13:37.858777 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/7d8475bc-7ad5-4381-b5dd-b45a6b9f73e2-installation-pull-secrets\") pod \"image-registry-66df7c8f76-z74gr\" (UID: \"7d8475bc-7ad5-4381-b5dd-b45a6b9f73e2\") " pod="openshift-image-registry/image-registry-66df7c8f76-z74gr" Dec 03 00:13:37 crc kubenswrapper[4953]: I1203 00:13:37.858854 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/7d8475bc-7ad5-4381-b5dd-b45a6b9f73e2-ca-trust-extracted\") pod \"image-registry-66df7c8f76-z74gr\" (UID: \"7d8475bc-7ad5-4381-b5dd-b45a6b9f73e2\") " pod="openshift-image-registry/image-registry-66df7c8f76-z74gr" Dec 03 00:13:37 crc kubenswrapper[4953]: I1203 00:13:37.858880 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/7d8475bc-7ad5-4381-b5dd-b45a6b9f73e2-trusted-ca\") pod \"image-registry-66df7c8f76-z74gr\" (UID: \"7d8475bc-7ad5-4381-b5dd-b45a6b9f73e2\") " pod="openshift-image-registry/image-registry-66df7c8f76-z74gr" Dec 03 00:13:37 crc kubenswrapper[4953]: I1203 00:13:37.858904 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/7d8475bc-7ad5-4381-b5dd-b45a6b9f73e2-registry-tls\") pod \"image-registry-66df7c8f76-z74gr\" (UID: \"7d8475bc-7ad5-4381-b5dd-b45a6b9f73e2\") " pod="openshift-image-registry/image-registry-66df7c8f76-z74gr" Dec 03 00:13:37 crc kubenswrapper[4953]: I1203 00:13:37.858933 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/7d8475bc-7ad5-4381-b5dd-b45a6b9f73e2-registry-certificates\") pod \"image-registry-66df7c8f76-z74gr\" (UID: \"7d8475bc-7ad5-4381-b5dd-b45a6b9f73e2\") " pod="openshift-image-registry/image-registry-66df7c8f76-z74gr" Dec 03 00:13:37 crc kubenswrapper[4953]: I1203 00:13:37.858954 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ct867\" (UniqueName: \"kubernetes.io/projected/7d8475bc-7ad5-4381-b5dd-b45a6b9f73e2-kube-api-access-ct867\") pod \"image-registry-66df7c8f76-z74gr\" (UID: \"7d8475bc-7ad5-4381-b5dd-b45a6b9f73e2\") " pod="openshift-image-registry/image-registry-66df7c8f76-z74gr" Dec 03 00:13:37 crc kubenswrapper[4953]: I1203 00:13:37.859069 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/7d8475bc-7ad5-4381-b5dd-b45a6b9f73e2-bound-sa-token\") pod \"image-registry-66df7c8f76-z74gr\" (UID: \"7d8475bc-7ad5-4381-b5dd-b45a6b9f73e2\") " pod="openshift-image-registry/image-registry-66df7c8f76-z74gr" Dec 03 00:13:37 crc kubenswrapper[4953]: I1203 00:13:37.859809 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/7d8475bc-7ad5-4381-b5dd-b45a6b9f73e2-ca-trust-extracted\") pod \"image-registry-66df7c8f76-z74gr\" (UID: \"7d8475bc-7ad5-4381-b5dd-b45a6b9f73e2\") " pod="openshift-image-registry/image-registry-66df7c8f76-z74gr" Dec 03 00:13:37 crc kubenswrapper[4953]: I1203 00:13:37.861177 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/7d8475bc-7ad5-4381-b5dd-b45a6b9f73e2-registry-certificates\") pod \"image-registry-66df7c8f76-z74gr\" (UID: \"7d8475bc-7ad5-4381-b5dd-b45a6b9f73e2\") " pod="openshift-image-registry/image-registry-66df7c8f76-z74gr" Dec 03 00:13:37 crc kubenswrapper[4953]: I1203 00:13:37.861316 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/7d8475bc-7ad5-4381-b5dd-b45a6b9f73e2-trusted-ca\") pod \"image-registry-66df7c8f76-z74gr\" (UID: \"7d8475bc-7ad5-4381-b5dd-b45a6b9f73e2\") " pod="openshift-image-registry/image-registry-66df7c8f76-z74gr" Dec 03 00:13:37 crc kubenswrapper[4953]: I1203 00:13:37.867216 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/7d8475bc-7ad5-4381-b5dd-b45a6b9f73e2-installation-pull-secrets\") pod \"image-registry-66df7c8f76-z74gr\" (UID: \"7d8475bc-7ad5-4381-b5dd-b45a6b9f73e2\") " pod="openshift-image-registry/image-registry-66df7c8f76-z74gr" Dec 03 00:13:37 crc kubenswrapper[4953]: I1203 00:13:37.869552 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/7d8475bc-7ad5-4381-b5dd-b45a6b9f73e2-registry-tls\") pod \"image-registry-66df7c8f76-z74gr\" (UID: \"7d8475bc-7ad5-4381-b5dd-b45a6b9f73e2\") " pod="openshift-image-registry/image-registry-66df7c8f76-z74gr" Dec 03 00:13:37 crc kubenswrapper[4953]: I1203 00:13:37.897156 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ct867\" (UniqueName: \"kubernetes.io/projected/7d8475bc-7ad5-4381-b5dd-b45a6b9f73e2-kube-api-access-ct867\") pod \"image-registry-66df7c8f76-z74gr\" (UID: \"7d8475bc-7ad5-4381-b5dd-b45a6b9f73e2\") " pod="openshift-image-registry/image-registry-66df7c8f76-z74gr" Dec 03 00:13:37 crc kubenswrapper[4953]: I1203 00:13:37.901456 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-556b846d59-5cf9d" event={"ID":"f677c67f-7084-4e3a-9255-404b4dcd8311","Type":"ContainerStarted","Data":"b928966cb2b299c64fde4b130fafa1e1a11e61f4c8589d0a2a70c5f02bbce65e"} Dec 03 00:13:37 crc kubenswrapper[4953]: I1203 00:13:37.902372 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/7d8475bc-7ad5-4381-b5dd-b45a6b9f73e2-bound-sa-token\") pod \"image-registry-66df7c8f76-z74gr\" (UID: \"7d8475bc-7ad5-4381-b5dd-b45a6b9f73e2\") " pod="openshift-image-registry/image-registry-66df7c8f76-z74gr" Dec 03 00:13:37 crc kubenswrapper[4953]: I1203 00:13:37.950098 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-z74gr" Dec 03 00:13:38 crc kubenswrapper[4953]: I1203 00:13:38.179718 4953 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-z74gr"] Dec 03 00:13:38 crc kubenswrapper[4953]: W1203 00:13:38.188343 4953 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7d8475bc_7ad5_4381_b5dd_b45a6b9f73e2.slice/crio-053e1a9d404128f8ac0da0963256ad32f87bdc1a1882031fccd8b055e7de4372 WatchSource:0}: Error finding container 053e1a9d404128f8ac0da0963256ad32f87bdc1a1882031fccd8b055e7de4372: Status 404 returned error can't find the container with id 053e1a9d404128f8ac0da0963256ad32f87bdc1a1882031fccd8b055e7de4372 Dec 03 00:13:38 crc kubenswrapper[4953]: I1203 00:13:38.910776 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-z74gr" event={"ID":"7d8475bc-7ad5-4381-b5dd-b45a6b9f73e2","Type":"ContainerStarted","Data":"149fde42e189e301c98fd9e2a5dc247373a1a30402d2aa9e9bd161f9bc2a5ed5"} Dec 03 00:13:38 crc kubenswrapper[4953]: I1203 00:13:38.911442 4953 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-66df7c8f76-z74gr" Dec 03 00:13:38 crc kubenswrapper[4953]: I1203 00:13:38.911459 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-z74gr" event={"ID":"7d8475bc-7ad5-4381-b5dd-b45a6b9f73e2","Type":"ContainerStarted","Data":"053e1a9d404128f8ac0da0963256ad32f87bdc1a1882031fccd8b055e7de4372"} Dec 03 00:13:38 crc kubenswrapper[4953]: I1203 00:13:38.912653 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-556b846d59-5cf9d" event={"ID":"f677c67f-7084-4e3a-9255-404b4dcd8311","Type":"ContainerStarted","Data":"2c050fdeead48fca14ae9c12759c3fe24da979bd579926021fc0941e4c1630e9"} Dec 03 00:13:38 crc kubenswrapper[4953]: I1203 00:13:38.913032 4953 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-556b846d59-5cf9d" Dec 03 00:13:38 crc kubenswrapper[4953]: I1203 00:13:38.919651 4953 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-556b846d59-5cf9d" Dec 03 00:13:38 crc kubenswrapper[4953]: I1203 00:13:38.937687 4953 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-66df7c8f76-z74gr" podStartSLOduration=1.9376654690000001 podStartE2EDuration="1.937665469s" podCreationTimestamp="2025-12-03 00:13:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 00:13:38.933495154 +0000 UTC m=+407.917775927" watchObservedRunningTime="2025-12-03 00:13:38.937665469 +0000 UTC m=+407.921946242" Dec 03 00:13:47 crc kubenswrapper[4953]: I1203 00:13:47.121119 4953 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-556b846d59-5cf9d" podStartSLOduration=12.121093188 podStartE2EDuration="12.121093188s" podCreationTimestamp="2025-12-03 00:13:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 00:13:38.957411376 +0000 UTC m=+407.941692159" watchObservedRunningTime="2025-12-03 00:13:47.121093188 +0000 UTC m=+416.105373961" Dec 03 00:13:47 crc kubenswrapper[4953]: I1203 00:13:47.126847 4953 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-xfd2l"] Dec 03 00:13:47 crc kubenswrapper[4953]: I1203 00:13:47.127194 4953 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-xfd2l" podUID="52fd5773-3888-49f1-813d-10d262f4c901" containerName="registry-server" containerID="cri-o://07abd4b77b78878cb3704fdc549bbd9b8f175c5bb93a78c71f5b651ea9c6e2f8" gracePeriod=30 Dec 03 00:13:47 crc kubenswrapper[4953]: I1203 00:13:47.152797 4953 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-hs6pg"] Dec 03 00:13:47 crc kubenswrapper[4953]: I1203 00:13:47.153241 4953 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-hs6pg" podUID="6b353eef-b7f9-4dbc-b2aa-9860220438cc" containerName="registry-server" containerID="cri-o://8af485ffdb0db0916a7f4f533b0b05a77458193f57ac4514e40e6bfc8205532f" gracePeriod=30 Dec 03 00:13:47 crc kubenswrapper[4953]: I1203 00:13:47.164136 4953 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-kx9dt"] Dec 03 00:13:47 crc kubenswrapper[4953]: I1203 00:13:47.164538 4953 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/marketplace-operator-79b997595-kx9dt" podUID="7124b81d-b87c-45de-a112-20602c4d9adb" containerName="marketplace-operator" containerID="cri-o://9a1303fc54264d3be3fb57969f4e7934b0ab3a22cc45c5e3c658bd67035f6751" gracePeriod=30 Dec 03 00:13:47 crc kubenswrapper[4953]: I1203 00:13:47.198550 4953 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-5zxsj"] Dec 03 00:13:47 crc kubenswrapper[4953]: I1203 00:13:47.199475 4953 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-5zxsj" podUID="e98b8740-f6fc-4db0-a807-664cf085391c" containerName="registry-server" containerID="cri-o://2bfe87b69385d5610e7e2c6550a4140ae496768d5035f606a6264aad12cc44e5" gracePeriod=30 Dec 03 00:13:47 crc kubenswrapper[4953]: I1203 00:13:47.204828 4953 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-dq6nn"] Dec 03 00:13:47 crc kubenswrapper[4953]: I1203 00:13:47.214345 4953 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-dq6nn" podUID="9064a5ab-24ee-4258-8363-6ce4ffa54b3f" containerName="registry-server" containerID="cri-o://d585b7d4d02d5b26ee4092ba449d6de3aa3d6e10bf9bc47f915048d6497ea65e" gracePeriod=30 Dec 03 00:13:47 crc kubenswrapper[4953]: I1203 00:13:47.260576 4953 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-d9khm"] Dec 03 00:13:47 crc kubenswrapper[4953]: I1203 00:13:47.261908 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-d9khm" Dec 03 00:13:47 crc kubenswrapper[4953]: I1203 00:13:47.277350 4953 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-d9khm"] Dec 03 00:13:47 crc kubenswrapper[4953]: I1203 00:13:47.378383 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/4ac535b1-a7ec-4999-aae4-d1c60c284aff-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-d9khm\" (UID: \"4ac535b1-a7ec-4999-aae4-d1c60c284aff\") " pod="openshift-marketplace/marketplace-operator-79b997595-d9khm" Dec 03 00:13:47 crc kubenswrapper[4953]: I1203 00:13:47.379208 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8hx4z\" (UniqueName: \"kubernetes.io/projected/4ac535b1-a7ec-4999-aae4-d1c60c284aff-kube-api-access-8hx4z\") pod \"marketplace-operator-79b997595-d9khm\" (UID: \"4ac535b1-a7ec-4999-aae4-d1c60c284aff\") " pod="openshift-marketplace/marketplace-operator-79b997595-d9khm" Dec 03 00:13:47 crc kubenswrapper[4953]: I1203 00:13:47.379433 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/4ac535b1-a7ec-4999-aae4-d1c60c284aff-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-d9khm\" (UID: \"4ac535b1-a7ec-4999-aae4-d1c60c284aff\") " pod="openshift-marketplace/marketplace-operator-79b997595-d9khm" Dec 03 00:13:47 crc kubenswrapper[4953]: I1203 00:13:47.481225 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/4ac535b1-a7ec-4999-aae4-d1c60c284aff-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-d9khm\" (UID: \"4ac535b1-a7ec-4999-aae4-d1c60c284aff\") " pod="openshift-marketplace/marketplace-operator-79b997595-d9khm" Dec 03 00:13:47 crc kubenswrapper[4953]: I1203 00:13:47.481292 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8hx4z\" (UniqueName: \"kubernetes.io/projected/4ac535b1-a7ec-4999-aae4-d1c60c284aff-kube-api-access-8hx4z\") pod \"marketplace-operator-79b997595-d9khm\" (UID: \"4ac535b1-a7ec-4999-aae4-d1c60c284aff\") " pod="openshift-marketplace/marketplace-operator-79b997595-d9khm" Dec 03 00:13:47 crc kubenswrapper[4953]: I1203 00:13:47.481319 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/4ac535b1-a7ec-4999-aae4-d1c60c284aff-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-d9khm\" (UID: \"4ac535b1-a7ec-4999-aae4-d1c60c284aff\") " pod="openshift-marketplace/marketplace-operator-79b997595-d9khm" Dec 03 00:13:47 crc kubenswrapper[4953]: I1203 00:13:47.483039 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/4ac535b1-a7ec-4999-aae4-d1c60c284aff-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-d9khm\" (UID: \"4ac535b1-a7ec-4999-aae4-d1c60c284aff\") " pod="openshift-marketplace/marketplace-operator-79b997595-d9khm" Dec 03 00:13:47 crc kubenswrapper[4953]: I1203 00:13:47.492153 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/4ac535b1-a7ec-4999-aae4-d1c60c284aff-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-d9khm\" (UID: \"4ac535b1-a7ec-4999-aae4-d1c60c284aff\") " pod="openshift-marketplace/marketplace-operator-79b997595-d9khm" Dec 03 00:13:47 crc kubenswrapper[4953]: I1203 00:13:47.503050 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8hx4z\" (UniqueName: \"kubernetes.io/projected/4ac535b1-a7ec-4999-aae4-d1c60c284aff-kube-api-access-8hx4z\") pod \"marketplace-operator-79b997595-d9khm\" (UID: \"4ac535b1-a7ec-4999-aae4-d1c60c284aff\") " pod="openshift-marketplace/marketplace-operator-79b997595-d9khm" Dec 03 00:13:47 crc kubenswrapper[4953]: I1203 00:13:47.634900 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-d9khm" Dec 03 00:13:47 crc kubenswrapper[4953]: I1203 00:13:47.867641 4953 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-5zxsj" Dec 03 00:13:47 crc kubenswrapper[4953]: I1203 00:13:47.948660 4953 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-dq6nn" Dec 03 00:13:47 crc kubenswrapper[4953]: I1203 00:13:47.988297 4953 generic.go:334] "Generic (PLEG): container finished" podID="52fd5773-3888-49f1-813d-10d262f4c901" containerID="07abd4b77b78878cb3704fdc549bbd9b8f175c5bb93a78c71f5b651ea9c6e2f8" exitCode=0 Dec 03 00:13:47 crc kubenswrapper[4953]: I1203 00:13:47.988409 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xfd2l" event={"ID":"52fd5773-3888-49f1-813d-10d262f4c901","Type":"ContainerDied","Data":"07abd4b77b78878cb3704fdc549bbd9b8f175c5bb93a78c71f5b651ea9c6e2f8"} Dec 03 00:13:47 crc kubenswrapper[4953]: I1203 00:13:47.990932 4953 generic.go:334] "Generic (PLEG): container finished" podID="9064a5ab-24ee-4258-8363-6ce4ffa54b3f" containerID="d585b7d4d02d5b26ee4092ba449d6de3aa3d6e10bf9bc47f915048d6497ea65e" exitCode=0 Dec 03 00:13:47 crc kubenswrapper[4953]: I1203 00:13:47.991018 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dq6nn" event={"ID":"9064a5ab-24ee-4258-8363-6ce4ffa54b3f","Type":"ContainerDied","Data":"d585b7d4d02d5b26ee4092ba449d6de3aa3d6e10bf9bc47f915048d6497ea65e"} Dec 03 00:13:47 crc kubenswrapper[4953]: I1203 00:13:47.991039 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dq6nn" event={"ID":"9064a5ab-24ee-4258-8363-6ce4ffa54b3f","Type":"ContainerDied","Data":"68ef5fb61b5fb8ed919f648e04de0009daebd9aed25ac57d75fe8c223f624424"} Dec 03 00:13:47 crc kubenswrapper[4953]: I1203 00:13:47.991063 4953 scope.go:117] "RemoveContainer" containerID="d585b7d4d02d5b26ee4092ba449d6de3aa3d6e10bf9bc47f915048d6497ea65e" Dec 03 00:13:47 crc kubenswrapper[4953]: I1203 00:13:47.991286 4953 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-dq6nn" Dec 03 00:13:47 crc kubenswrapper[4953]: I1203 00:13:47.998567 4953 generic.go:334] "Generic (PLEG): container finished" podID="6b353eef-b7f9-4dbc-b2aa-9860220438cc" containerID="8af485ffdb0db0916a7f4f533b0b05a77458193f57ac4514e40e6bfc8205532f" exitCode=0 Dec 03 00:13:47 crc kubenswrapper[4953]: I1203 00:13:47.998638 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hs6pg" event={"ID":"6b353eef-b7f9-4dbc-b2aa-9860220438cc","Type":"ContainerDied","Data":"8af485ffdb0db0916a7f4f533b0b05a77458193f57ac4514e40e6bfc8205532f"} Dec 03 00:13:48 crc kubenswrapper[4953]: I1203 00:13:48.001435 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-kx9dt" event={"ID":"7124b81d-b87c-45de-a112-20602c4d9adb","Type":"ContainerDied","Data":"9a1303fc54264d3be3fb57969f4e7934b0ab3a22cc45c5e3c658bd67035f6751"} Dec 03 00:13:48 crc kubenswrapper[4953]: I1203 00:13:48.001455 4953 generic.go:334] "Generic (PLEG): container finished" podID="7124b81d-b87c-45de-a112-20602c4d9adb" containerID="9a1303fc54264d3be3fb57969f4e7934b0ab3a22cc45c5e3c658bd67035f6751" exitCode=0 Dec 03 00:13:48 crc kubenswrapper[4953]: I1203 00:13:48.004707 4953 generic.go:334] "Generic (PLEG): container finished" podID="e98b8740-f6fc-4db0-a807-664cf085391c" containerID="2bfe87b69385d5610e7e2c6550a4140ae496768d5035f606a6264aad12cc44e5" exitCode=0 Dec 03 00:13:48 crc kubenswrapper[4953]: I1203 00:13:48.004775 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5zxsj" event={"ID":"e98b8740-f6fc-4db0-a807-664cf085391c","Type":"ContainerDied","Data":"2bfe87b69385d5610e7e2c6550a4140ae496768d5035f606a6264aad12cc44e5"} Dec 03 00:13:48 crc kubenswrapper[4953]: I1203 00:13:48.004882 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5zxsj" event={"ID":"e98b8740-f6fc-4db0-a807-664cf085391c","Type":"ContainerDied","Data":"c747d83b368b5a45d4daead722975738b091c4d0e85f6a787fc22c844d1bd7dd"} Dec 03 00:13:48 crc kubenswrapper[4953]: I1203 00:13:48.005015 4953 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-5zxsj" Dec 03 00:13:48 crc kubenswrapper[4953]: I1203 00:13:48.006034 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e98b8740-f6fc-4db0-a807-664cf085391c-utilities\") pod \"e98b8740-f6fc-4db0-a807-664cf085391c\" (UID: \"e98b8740-f6fc-4db0-a807-664cf085391c\") " Dec 03 00:13:48 crc kubenswrapper[4953]: I1203 00:13:48.006235 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e98b8740-f6fc-4db0-a807-664cf085391c-catalog-content\") pod \"e98b8740-f6fc-4db0-a807-664cf085391c\" (UID: \"e98b8740-f6fc-4db0-a807-664cf085391c\") " Dec 03 00:13:48 crc kubenswrapper[4953]: I1203 00:13:48.006406 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kv2jf\" (UniqueName: \"kubernetes.io/projected/e98b8740-f6fc-4db0-a807-664cf085391c-kube-api-access-kv2jf\") pod \"e98b8740-f6fc-4db0-a807-664cf085391c\" (UID: \"e98b8740-f6fc-4db0-a807-664cf085391c\") " Dec 03 00:13:48 crc kubenswrapper[4953]: I1203 00:13:48.008013 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e98b8740-f6fc-4db0-a807-664cf085391c-utilities" (OuterVolumeSpecName: "utilities") pod "e98b8740-f6fc-4db0-a807-664cf085391c" (UID: "e98b8740-f6fc-4db0-a807-664cf085391c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 00:13:48 crc kubenswrapper[4953]: I1203 00:13:48.016256 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e98b8740-f6fc-4db0-a807-664cf085391c-kube-api-access-kv2jf" (OuterVolumeSpecName: "kube-api-access-kv2jf") pod "e98b8740-f6fc-4db0-a807-664cf085391c" (UID: "e98b8740-f6fc-4db0-a807-664cf085391c"). InnerVolumeSpecName "kube-api-access-kv2jf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:13:48 crc kubenswrapper[4953]: I1203 00:13:48.038125 4953 scope.go:117] "RemoveContainer" containerID="5409b03d52f75ac22e98a311685907066f3577a6782501f328e36f36f67641c7" Dec 03 00:13:48 crc kubenswrapper[4953]: I1203 00:13:48.039679 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e98b8740-f6fc-4db0-a807-664cf085391c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e98b8740-f6fc-4db0-a807-664cf085391c" (UID: "e98b8740-f6fc-4db0-a807-664cf085391c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 00:13:48 crc kubenswrapper[4953]: I1203 00:13:48.058556 4953 scope.go:117] "RemoveContainer" containerID="beeb25d3d42d4d66ef7fae9fc6aa9b0ae72ab4540161cf231a11bd225d262597" Dec 03 00:13:48 crc kubenswrapper[4953]: I1203 00:13:48.078002 4953 scope.go:117] "RemoveContainer" containerID="d585b7d4d02d5b26ee4092ba449d6de3aa3d6e10bf9bc47f915048d6497ea65e" Dec 03 00:13:48 crc kubenswrapper[4953]: E1203 00:13:48.079961 4953 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d585b7d4d02d5b26ee4092ba449d6de3aa3d6e10bf9bc47f915048d6497ea65e\": container with ID starting with d585b7d4d02d5b26ee4092ba449d6de3aa3d6e10bf9bc47f915048d6497ea65e not found: ID does not exist" containerID="d585b7d4d02d5b26ee4092ba449d6de3aa3d6e10bf9bc47f915048d6497ea65e" Dec 03 00:13:48 crc kubenswrapper[4953]: I1203 00:13:48.080031 4953 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d585b7d4d02d5b26ee4092ba449d6de3aa3d6e10bf9bc47f915048d6497ea65e"} err="failed to get container status \"d585b7d4d02d5b26ee4092ba449d6de3aa3d6e10bf9bc47f915048d6497ea65e\": rpc error: code = NotFound desc = could not find container \"d585b7d4d02d5b26ee4092ba449d6de3aa3d6e10bf9bc47f915048d6497ea65e\": container with ID starting with d585b7d4d02d5b26ee4092ba449d6de3aa3d6e10bf9bc47f915048d6497ea65e not found: ID does not exist" Dec 03 00:13:48 crc kubenswrapper[4953]: I1203 00:13:48.080065 4953 scope.go:117] "RemoveContainer" containerID="5409b03d52f75ac22e98a311685907066f3577a6782501f328e36f36f67641c7" Dec 03 00:13:48 crc kubenswrapper[4953]: E1203 00:13:48.080883 4953 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5409b03d52f75ac22e98a311685907066f3577a6782501f328e36f36f67641c7\": container with ID starting with 5409b03d52f75ac22e98a311685907066f3577a6782501f328e36f36f67641c7 not found: ID does not exist" containerID="5409b03d52f75ac22e98a311685907066f3577a6782501f328e36f36f67641c7" Dec 03 00:13:48 crc kubenswrapper[4953]: I1203 00:13:48.080917 4953 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5409b03d52f75ac22e98a311685907066f3577a6782501f328e36f36f67641c7"} err="failed to get container status \"5409b03d52f75ac22e98a311685907066f3577a6782501f328e36f36f67641c7\": rpc error: code = NotFound desc = could not find container \"5409b03d52f75ac22e98a311685907066f3577a6782501f328e36f36f67641c7\": container with ID starting with 5409b03d52f75ac22e98a311685907066f3577a6782501f328e36f36f67641c7 not found: ID does not exist" Dec 03 00:13:48 crc kubenswrapper[4953]: I1203 00:13:48.080939 4953 scope.go:117] "RemoveContainer" containerID="beeb25d3d42d4d66ef7fae9fc6aa9b0ae72ab4540161cf231a11bd225d262597" Dec 03 00:13:48 crc kubenswrapper[4953]: E1203 00:13:48.081513 4953 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"beeb25d3d42d4d66ef7fae9fc6aa9b0ae72ab4540161cf231a11bd225d262597\": container with ID starting with beeb25d3d42d4d66ef7fae9fc6aa9b0ae72ab4540161cf231a11bd225d262597 not found: ID does not exist" containerID="beeb25d3d42d4d66ef7fae9fc6aa9b0ae72ab4540161cf231a11bd225d262597" Dec 03 00:13:48 crc kubenswrapper[4953]: I1203 00:13:48.081535 4953 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"beeb25d3d42d4d66ef7fae9fc6aa9b0ae72ab4540161cf231a11bd225d262597"} err="failed to get container status \"beeb25d3d42d4d66ef7fae9fc6aa9b0ae72ab4540161cf231a11bd225d262597\": rpc error: code = NotFound desc = could not find container \"beeb25d3d42d4d66ef7fae9fc6aa9b0ae72ab4540161cf231a11bd225d262597\": container with ID starting with beeb25d3d42d4d66ef7fae9fc6aa9b0ae72ab4540161cf231a11bd225d262597 not found: ID does not exist" Dec 03 00:13:48 crc kubenswrapper[4953]: I1203 00:13:48.081553 4953 scope.go:117] "RemoveContainer" containerID="cf37583f0764dfdbf86f5e9f0dae7b270a80910c7793261d30e36b161b46cea5" Dec 03 00:13:48 crc kubenswrapper[4953]: I1203 00:13:48.108142 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9064a5ab-24ee-4258-8363-6ce4ffa54b3f-catalog-content\") pod \"9064a5ab-24ee-4258-8363-6ce4ffa54b3f\" (UID: \"9064a5ab-24ee-4258-8363-6ce4ffa54b3f\") " Dec 03 00:13:48 crc kubenswrapper[4953]: I1203 00:13:48.108248 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9064a5ab-24ee-4258-8363-6ce4ffa54b3f-utilities\") pod \"9064a5ab-24ee-4258-8363-6ce4ffa54b3f\" (UID: \"9064a5ab-24ee-4258-8363-6ce4ffa54b3f\") " Dec 03 00:13:48 crc kubenswrapper[4953]: I1203 00:13:48.108326 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8d5tx\" (UniqueName: \"kubernetes.io/projected/9064a5ab-24ee-4258-8363-6ce4ffa54b3f-kube-api-access-8d5tx\") pod \"9064a5ab-24ee-4258-8363-6ce4ffa54b3f\" (UID: \"9064a5ab-24ee-4258-8363-6ce4ffa54b3f\") " Dec 03 00:13:48 crc kubenswrapper[4953]: I1203 00:13:48.108658 4953 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e98b8740-f6fc-4db0-a807-664cf085391c-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 00:13:48 crc kubenswrapper[4953]: I1203 00:13:48.108677 4953 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e98b8740-f6fc-4db0-a807-664cf085391c-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 00:13:48 crc kubenswrapper[4953]: I1203 00:13:48.108688 4953 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kv2jf\" (UniqueName: \"kubernetes.io/projected/e98b8740-f6fc-4db0-a807-664cf085391c-kube-api-access-kv2jf\") on node \"crc\" DevicePath \"\"" Dec 03 00:13:48 crc kubenswrapper[4953]: I1203 00:13:48.110745 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9064a5ab-24ee-4258-8363-6ce4ffa54b3f-utilities" (OuterVolumeSpecName: "utilities") pod "9064a5ab-24ee-4258-8363-6ce4ffa54b3f" (UID: "9064a5ab-24ee-4258-8363-6ce4ffa54b3f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 00:13:48 crc kubenswrapper[4953]: I1203 00:13:48.117490 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9064a5ab-24ee-4258-8363-6ce4ffa54b3f-kube-api-access-8d5tx" (OuterVolumeSpecName: "kube-api-access-8d5tx") pod "9064a5ab-24ee-4258-8363-6ce4ffa54b3f" (UID: "9064a5ab-24ee-4258-8363-6ce4ffa54b3f"). InnerVolumeSpecName "kube-api-access-8d5tx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:13:48 crc kubenswrapper[4953]: I1203 00:13:48.142958 4953 scope.go:117] "RemoveContainer" containerID="2bfe87b69385d5610e7e2c6550a4140ae496768d5035f606a6264aad12cc44e5" Dec 03 00:13:48 crc kubenswrapper[4953]: I1203 00:13:48.168364 4953 scope.go:117] "RemoveContainer" containerID="ab8f2200b1995da972fc62acce6b9433d30b466901c9ab10b23bbb9a4e2c4bf4" Dec 03 00:13:48 crc kubenswrapper[4953]: I1203 00:13:48.185464 4953 scope.go:117] "RemoveContainer" containerID="7033f14e4fa2cbc2578476d618efb3e44d904a9138d5b8ac8ec8d5d143af0f11" Dec 03 00:13:48 crc kubenswrapper[4953]: I1203 00:13:48.210435 4953 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8d5tx\" (UniqueName: \"kubernetes.io/projected/9064a5ab-24ee-4258-8363-6ce4ffa54b3f-kube-api-access-8d5tx\") on node \"crc\" DevicePath \"\"" Dec 03 00:13:48 crc kubenswrapper[4953]: I1203 00:13:48.210467 4953 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9064a5ab-24ee-4258-8363-6ce4ffa54b3f-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 00:13:48 crc kubenswrapper[4953]: I1203 00:13:48.230074 4953 scope.go:117] "RemoveContainer" containerID="2bfe87b69385d5610e7e2c6550a4140ae496768d5035f606a6264aad12cc44e5" Dec 03 00:13:48 crc kubenswrapper[4953]: E1203 00:13:48.233239 4953 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2bfe87b69385d5610e7e2c6550a4140ae496768d5035f606a6264aad12cc44e5\": container with ID starting with 2bfe87b69385d5610e7e2c6550a4140ae496768d5035f606a6264aad12cc44e5 not found: ID does not exist" containerID="2bfe87b69385d5610e7e2c6550a4140ae496768d5035f606a6264aad12cc44e5" Dec 03 00:13:48 crc kubenswrapper[4953]: I1203 00:13:48.233274 4953 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2bfe87b69385d5610e7e2c6550a4140ae496768d5035f606a6264aad12cc44e5"} err="failed to get container status \"2bfe87b69385d5610e7e2c6550a4140ae496768d5035f606a6264aad12cc44e5\": rpc error: code = NotFound desc = could not find container \"2bfe87b69385d5610e7e2c6550a4140ae496768d5035f606a6264aad12cc44e5\": container with ID starting with 2bfe87b69385d5610e7e2c6550a4140ae496768d5035f606a6264aad12cc44e5 not found: ID does not exist" Dec 03 00:13:48 crc kubenswrapper[4953]: I1203 00:13:48.233304 4953 scope.go:117] "RemoveContainer" containerID="ab8f2200b1995da972fc62acce6b9433d30b466901c9ab10b23bbb9a4e2c4bf4" Dec 03 00:13:48 crc kubenswrapper[4953]: E1203 00:13:48.233688 4953 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ab8f2200b1995da972fc62acce6b9433d30b466901c9ab10b23bbb9a4e2c4bf4\": container with ID starting with ab8f2200b1995da972fc62acce6b9433d30b466901c9ab10b23bbb9a4e2c4bf4 not found: ID does not exist" containerID="ab8f2200b1995da972fc62acce6b9433d30b466901c9ab10b23bbb9a4e2c4bf4" Dec 03 00:13:48 crc kubenswrapper[4953]: I1203 00:13:48.233706 4953 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ab8f2200b1995da972fc62acce6b9433d30b466901c9ab10b23bbb9a4e2c4bf4"} err="failed to get container status \"ab8f2200b1995da972fc62acce6b9433d30b466901c9ab10b23bbb9a4e2c4bf4\": rpc error: code = NotFound desc = could not find container \"ab8f2200b1995da972fc62acce6b9433d30b466901c9ab10b23bbb9a4e2c4bf4\": container with ID starting with ab8f2200b1995da972fc62acce6b9433d30b466901c9ab10b23bbb9a4e2c4bf4 not found: ID does not exist" Dec 03 00:13:48 crc kubenswrapper[4953]: I1203 00:13:48.233738 4953 scope.go:117] "RemoveContainer" containerID="7033f14e4fa2cbc2578476d618efb3e44d904a9138d5b8ac8ec8d5d143af0f11" Dec 03 00:13:48 crc kubenswrapper[4953]: E1203 00:13:48.234088 4953 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7033f14e4fa2cbc2578476d618efb3e44d904a9138d5b8ac8ec8d5d143af0f11\": container with ID starting with 7033f14e4fa2cbc2578476d618efb3e44d904a9138d5b8ac8ec8d5d143af0f11 not found: ID does not exist" containerID="7033f14e4fa2cbc2578476d618efb3e44d904a9138d5b8ac8ec8d5d143af0f11" Dec 03 00:13:48 crc kubenswrapper[4953]: I1203 00:13:48.234112 4953 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7033f14e4fa2cbc2578476d618efb3e44d904a9138d5b8ac8ec8d5d143af0f11"} err="failed to get container status \"7033f14e4fa2cbc2578476d618efb3e44d904a9138d5b8ac8ec8d5d143af0f11\": rpc error: code = NotFound desc = could not find container \"7033f14e4fa2cbc2578476d618efb3e44d904a9138d5b8ac8ec8d5d143af0f11\": container with ID starting with 7033f14e4fa2cbc2578476d618efb3e44d904a9138d5b8ac8ec8d5d143af0f11 not found: ID does not exist" Dec 03 00:13:48 crc kubenswrapper[4953]: I1203 00:13:48.245111 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9064a5ab-24ee-4258-8363-6ce4ffa54b3f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "9064a5ab-24ee-4258-8363-6ce4ffa54b3f" (UID: "9064a5ab-24ee-4258-8363-6ce4ffa54b3f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 00:13:48 crc kubenswrapper[4953]: I1203 00:13:48.311501 4953 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9064a5ab-24ee-4258-8363-6ce4ffa54b3f-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 00:13:48 crc kubenswrapper[4953]: I1203 00:13:48.328953 4953 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-dq6nn"] Dec 03 00:13:48 crc kubenswrapper[4953]: I1203 00:13:48.333355 4953 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-dq6nn"] Dec 03 00:13:48 crc kubenswrapper[4953]: I1203 00:13:48.350588 4953 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-5zxsj"] Dec 03 00:13:48 crc kubenswrapper[4953]: I1203 00:13:48.385983 4953 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-5zxsj"] Dec 03 00:13:48 crc kubenswrapper[4953]: I1203 00:13:48.409473 4953 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-d9khm"] Dec 03 00:13:48 crc kubenswrapper[4953]: I1203 00:13:48.586587 4953 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-xfd2l" Dec 03 00:13:48 crc kubenswrapper[4953]: I1203 00:13:48.604371 4953 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-kx9dt" Dec 03 00:13:48 crc kubenswrapper[4953]: I1203 00:13:48.609746 4953 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-hs6pg" Dec 03 00:13:48 crc kubenswrapper[4953]: I1203 00:13:48.719557 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5bm5b\" (UniqueName: \"kubernetes.io/projected/7124b81d-b87c-45de-a112-20602c4d9adb-kube-api-access-5bm5b\") pod \"7124b81d-b87c-45de-a112-20602c4d9adb\" (UID: \"7124b81d-b87c-45de-a112-20602c4d9adb\") " Dec 03 00:13:48 crc kubenswrapper[4953]: I1203 00:13:48.720185 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6b353eef-b7f9-4dbc-b2aa-9860220438cc-catalog-content\") pod \"6b353eef-b7f9-4dbc-b2aa-9860220438cc\" (UID: \"6b353eef-b7f9-4dbc-b2aa-9860220438cc\") " Dec 03 00:13:48 crc kubenswrapper[4953]: I1203 00:13:48.720286 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6b353eef-b7f9-4dbc-b2aa-9860220438cc-utilities\") pod \"6b353eef-b7f9-4dbc-b2aa-9860220438cc\" (UID: \"6b353eef-b7f9-4dbc-b2aa-9860220438cc\") " Dec 03 00:13:48 crc kubenswrapper[4953]: I1203 00:13:48.720309 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/52fd5773-3888-49f1-813d-10d262f4c901-utilities\") pod \"52fd5773-3888-49f1-813d-10d262f4c901\" (UID: \"52fd5773-3888-49f1-813d-10d262f4c901\") " Dec 03 00:13:48 crc kubenswrapper[4953]: I1203 00:13:48.720334 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/7124b81d-b87c-45de-a112-20602c4d9adb-marketplace-operator-metrics\") pod \"7124b81d-b87c-45de-a112-20602c4d9adb\" (UID: \"7124b81d-b87c-45de-a112-20602c4d9adb\") " Dec 03 00:13:48 crc kubenswrapper[4953]: I1203 00:13:48.720361 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m89m9\" (UniqueName: \"kubernetes.io/projected/6b353eef-b7f9-4dbc-b2aa-9860220438cc-kube-api-access-m89m9\") pod \"6b353eef-b7f9-4dbc-b2aa-9860220438cc\" (UID: \"6b353eef-b7f9-4dbc-b2aa-9860220438cc\") " Dec 03 00:13:48 crc kubenswrapper[4953]: I1203 00:13:48.720384 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/52fd5773-3888-49f1-813d-10d262f4c901-catalog-content\") pod \"52fd5773-3888-49f1-813d-10d262f4c901\" (UID: \"52fd5773-3888-49f1-813d-10d262f4c901\") " Dec 03 00:13:48 crc kubenswrapper[4953]: I1203 00:13:48.720415 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/7124b81d-b87c-45de-a112-20602c4d9adb-marketplace-trusted-ca\") pod \"7124b81d-b87c-45de-a112-20602c4d9adb\" (UID: \"7124b81d-b87c-45de-a112-20602c4d9adb\") " Dec 03 00:13:48 crc kubenswrapper[4953]: I1203 00:13:48.721270 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/52fd5773-3888-49f1-813d-10d262f4c901-utilities" (OuterVolumeSpecName: "utilities") pod "52fd5773-3888-49f1-813d-10d262f4c901" (UID: "52fd5773-3888-49f1-813d-10d262f4c901"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 00:13:48 crc kubenswrapper[4953]: I1203 00:13:48.721329 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6b353eef-b7f9-4dbc-b2aa-9860220438cc-utilities" (OuterVolumeSpecName: "utilities") pod "6b353eef-b7f9-4dbc-b2aa-9860220438cc" (UID: "6b353eef-b7f9-4dbc-b2aa-9860220438cc"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 00:13:48 crc kubenswrapper[4953]: I1203 00:13:48.721385 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7124b81d-b87c-45de-a112-20602c4d9adb-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "7124b81d-b87c-45de-a112-20602c4d9adb" (UID: "7124b81d-b87c-45de-a112-20602c4d9adb"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:13:48 crc kubenswrapper[4953]: I1203 00:13:48.721489 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qlmtg\" (UniqueName: \"kubernetes.io/projected/52fd5773-3888-49f1-813d-10d262f4c901-kube-api-access-qlmtg\") pod \"52fd5773-3888-49f1-813d-10d262f4c901\" (UID: \"52fd5773-3888-49f1-813d-10d262f4c901\") " Dec 03 00:13:48 crc kubenswrapper[4953]: I1203 00:13:48.724582 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7124b81d-b87c-45de-a112-20602c4d9adb-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "7124b81d-b87c-45de-a112-20602c4d9adb" (UID: "7124b81d-b87c-45de-a112-20602c4d9adb"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:13:48 crc kubenswrapper[4953]: I1203 00:13:48.725956 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/52fd5773-3888-49f1-813d-10d262f4c901-kube-api-access-qlmtg" (OuterVolumeSpecName: "kube-api-access-qlmtg") pod "52fd5773-3888-49f1-813d-10d262f4c901" (UID: "52fd5773-3888-49f1-813d-10d262f4c901"). InnerVolumeSpecName "kube-api-access-qlmtg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:13:48 crc kubenswrapper[4953]: I1203 00:13:48.726350 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7124b81d-b87c-45de-a112-20602c4d9adb-kube-api-access-5bm5b" (OuterVolumeSpecName: "kube-api-access-5bm5b") pod "7124b81d-b87c-45de-a112-20602c4d9adb" (UID: "7124b81d-b87c-45de-a112-20602c4d9adb"). InnerVolumeSpecName "kube-api-access-5bm5b". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:13:48 crc kubenswrapper[4953]: I1203 00:13:48.727917 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6b353eef-b7f9-4dbc-b2aa-9860220438cc-kube-api-access-m89m9" (OuterVolumeSpecName: "kube-api-access-m89m9") pod "6b353eef-b7f9-4dbc-b2aa-9860220438cc" (UID: "6b353eef-b7f9-4dbc-b2aa-9860220438cc"). InnerVolumeSpecName "kube-api-access-m89m9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:13:48 crc kubenswrapper[4953]: I1203 00:13:48.730799 4953 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5bm5b\" (UniqueName: \"kubernetes.io/projected/7124b81d-b87c-45de-a112-20602c4d9adb-kube-api-access-5bm5b\") on node \"crc\" DevicePath \"\"" Dec 03 00:13:48 crc kubenswrapper[4953]: I1203 00:13:48.730832 4953 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6b353eef-b7f9-4dbc-b2aa-9860220438cc-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 00:13:48 crc kubenswrapper[4953]: I1203 00:13:48.730844 4953 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/52fd5773-3888-49f1-813d-10d262f4c901-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 00:13:48 crc kubenswrapper[4953]: I1203 00:13:48.730859 4953 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/7124b81d-b87c-45de-a112-20602c4d9adb-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Dec 03 00:13:48 crc kubenswrapper[4953]: I1203 00:13:48.730872 4953 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m89m9\" (UniqueName: \"kubernetes.io/projected/6b353eef-b7f9-4dbc-b2aa-9860220438cc-kube-api-access-m89m9\") on node \"crc\" DevicePath \"\"" Dec 03 00:13:48 crc kubenswrapper[4953]: I1203 00:13:48.730883 4953 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/7124b81d-b87c-45de-a112-20602c4d9adb-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 03 00:13:48 crc kubenswrapper[4953]: I1203 00:13:48.730893 4953 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qlmtg\" (UniqueName: \"kubernetes.io/projected/52fd5773-3888-49f1-813d-10d262f4c901-kube-api-access-qlmtg\") on node \"crc\" DevicePath \"\"" Dec 03 00:13:48 crc kubenswrapper[4953]: I1203 00:13:48.778625 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/52fd5773-3888-49f1-813d-10d262f4c901-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "52fd5773-3888-49f1-813d-10d262f4c901" (UID: "52fd5773-3888-49f1-813d-10d262f4c901"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 00:13:48 crc kubenswrapper[4953]: I1203 00:13:48.784405 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6b353eef-b7f9-4dbc-b2aa-9860220438cc-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "6b353eef-b7f9-4dbc-b2aa-9860220438cc" (UID: "6b353eef-b7f9-4dbc-b2aa-9860220438cc"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 00:13:48 crc kubenswrapper[4953]: I1203 00:13:48.832718 4953 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6b353eef-b7f9-4dbc-b2aa-9860220438cc-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 00:13:48 crc kubenswrapper[4953]: I1203 00:13:48.832770 4953 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/52fd5773-3888-49f1-813d-10d262f4c901-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 00:13:48 crc kubenswrapper[4953]: I1203 00:13:48.945095 4953 patch_prober.go:28] interesting pod/machine-config-daemon-6dfbn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 00:13:48 crc kubenswrapper[4953]: I1203 00:13:48.945185 4953 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-6dfbn" podUID="53ffb1e9-9a2e-4869-80d7-776f4aa7ef8b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 00:13:48 crc kubenswrapper[4953]: I1203 00:13:48.945257 4953 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-6dfbn" Dec 03 00:13:48 crc kubenswrapper[4953]: I1203 00:13:48.946255 4953 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"cd3fdc5de406f82703cc48138f69373c50d1862f3fb1f4cc4572aa863d2eca4c"} pod="openshift-machine-config-operator/machine-config-daemon-6dfbn" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 03 00:13:48 crc kubenswrapper[4953]: I1203 00:13:48.946332 4953 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-6dfbn" podUID="53ffb1e9-9a2e-4869-80d7-776f4aa7ef8b" containerName="machine-config-daemon" containerID="cri-o://cd3fdc5de406f82703cc48138f69373c50d1862f3fb1f4cc4572aa863d2eca4c" gracePeriod=600 Dec 03 00:13:49 crc kubenswrapper[4953]: I1203 00:13:49.013705 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xfd2l" event={"ID":"52fd5773-3888-49f1-813d-10d262f4c901","Type":"ContainerDied","Data":"37d8cbc201645bf6b66467acb62ab2437ecd3b0b9d98aa4cfefee38db9f2b0ee"} Dec 03 00:13:49 crc kubenswrapper[4953]: I1203 00:13:49.013785 4953 scope.go:117] "RemoveContainer" containerID="07abd4b77b78878cb3704fdc549bbd9b8f175c5bb93a78c71f5b651ea9c6e2f8" Dec 03 00:13:49 crc kubenswrapper[4953]: I1203 00:13:49.013939 4953 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-xfd2l" Dec 03 00:13:49 crc kubenswrapper[4953]: I1203 00:13:49.020933 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-d9khm" event={"ID":"4ac535b1-a7ec-4999-aae4-d1c60c284aff","Type":"ContainerStarted","Data":"675b1bda01761f8c0c79f4a4f100efd0207ac7949073ddc73d1aac7299c634b0"} Dec 03 00:13:49 crc kubenswrapper[4953]: I1203 00:13:49.021019 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-d9khm" event={"ID":"4ac535b1-a7ec-4999-aae4-d1c60c284aff","Type":"ContainerStarted","Data":"507a8b496e725774d4e8141f23efc5bca9d561419945c3d67d87eec0f3ad6af2"} Dec 03 00:13:49 crc kubenswrapper[4953]: I1203 00:13:49.021224 4953 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-d9khm" Dec 03 00:13:49 crc kubenswrapper[4953]: I1203 00:13:49.023151 4953 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-kx9dt" Dec 03 00:13:49 crc kubenswrapper[4953]: I1203 00:13:49.023707 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-kx9dt" event={"ID":"7124b81d-b87c-45de-a112-20602c4d9adb","Type":"ContainerDied","Data":"f8da00cbcdc757fed1f6c9775d6af01029548aca767571414b5aa799e2566daf"} Dec 03 00:13:49 crc kubenswrapper[4953]: I1203 00:13:49.025006 4953 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-d9khm" Dec 03 00:13:49 crc kubenswrapper[4953]: I1203 00:13:49.028009 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hs6pg" event={"ID":"6b353eef-b7f9-4dbc-b2aa-9860220438cc","Type":"ContainerDied","Data":"224b30baa9aff1a09e27d6c314399ce59aee5ea2b5e60b153e20cbadcb677515"} Dec 03 00:13:49 crc kubenswrapper[4953]: I1203 00:13:49.028080 4953 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-hs6pg" Dec 03 00:13:49 crc kubenswrapper[4953]: I1203 00:13:49.034274 4953 scope.go:117] "RemoveContainer" containerID="3a3bd25c8318f951154af34ed59bdf755743e6f4b6d7bc248b6828f0b46d7888" Dec 03 00:13:49 crc kubenswrapper[4953]: I1203 00:13:49.066499 4953 scope.go:117] "RemoveContainer" containerID="a9095f2983a7ed1cb660c3459aaa9888337df9b6cd3750a5a83a8ccf53a7a70c" Dec 03 00:13:49 crc kubenswrapper[4953]: I1203 00:13:49.092354 4953 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-d9khm" podStartSLOduration=2.092319621 podStartE2EDuration="2.092319621s" podCreationTimestamp="2025-12-03 00:13:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 00:13:49.059741812 +0000 UTC m=+418.044022595" watchObservedRunningTime="2025-12-03 00:13:49.092319621 +0000 UTC m=+418.076600405" Dec 03 00:13:49 crc kubenswrapper[4953]: I1203 00:13:49.093896 4953 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-xfd2l"] Dec 03 00:13:49 crc kubenswrapper[4953]: I1203 00:13:49.112192 4953 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-xfd2l"] Dec 03 00:13:49 crc kubenswrapper[4953]: I1203 00:13:49.116220 4953 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-kx9dt"] Dec 03 00:13:49 crc kubenswrapper[4953]: I1203 00:13:49.117808 4953 scope.go:117] "RemoveContainer" containerID="9a1303fc54264d3be3fb57969f4e7934b0ab3a22cc45c5e3c658bd67035f6751" Dec 03 00:13:49 crc kubenswrapper[4953]: I1203 00:13:49.132136 4953 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-kx9dt"] Dec 03 00:13:49 crc kubenswrapper[4953]: I1203 00:13:49.138037 4953 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-hs6pg"] Dec 03 00:13:49 crc kubenswrapper[4953]: I1203 00:13:49.143318 4953 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-hs6pg"] Dec 03 00:13:49 crc kubenswrapper[4953]: I1203 00:13:49.152116 4953 scope.go:117] "RemoveContainer" containerID="8af485ffdb0db0916a7f4f533b0b05a77458193f57ac4514e40e6bfc8205532f" Dec 03 00:13:49 crc kubenswrapper[4953]: I1203 00:13:49.174831 4953 scope.go:117] "RemoveContainer" containerID="b47d72f0f01ce88a842fca13012bced12a6ffb7e74cf75e86142b1f1c4300e80" Dec 03 00:13:49 crc kubenswrapper[4953]: I1203 00:13:49.205650 4953 scope.go:117] "RemoveContainer" containerID="261b04a7d566d558fe6b4892804ef231ee5a882005acbe560446cd4993e5fc9c" Dec 03 00:13:49 crc kubenswrapper[4953]: I1203 00:13:49.361291 4953 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="52fd5773-3888-49f1-813d-10d262f4c901" path="/var/lib/kubelet/pods/52fd5773-3888-49f1-813d-10d262f4c901/volumes" Dec 03 00:13:49 crc kubenswrapper[4953]: I1203 00:13:49.362200 4953 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6b353eef-b7f9-4dbc-b2aa-9860220438cc" path="/var/lib/kubelet/pods/6b353eef-b7f9-4dbc-b2aa-9860220438cc/volumes" Dec 03 00:13:49 crc kubenswrapper[4953]: I1203 00:13:49.363126 4953 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7124b81d-b87c-45de-a112-20602c4d9adb" path="/var/lib/kubelet/pods/7124b81d-b87c-45de-a112-20602c4d9adb/volumes" Dec 03 00:13:49 crc kubenswrapper[4953]: I1203 00:13:49.364488 4953 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9064a5ab-24ee-4258-8363-6ce4ffa54b3f" path="/var/lib/kubelet/pods/9064a5ab-24ee-4258-8363-6ce4ffa54b3f/volumes" Dec 03 00:13:49 crc kubenswrapper[4953]: I1203 00:13:49.365299 4953 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e98b8740-f6fc-4db0-a807-664cf085391c" path="/var/lib/kubelet/pods/e98b8740-f6fc-4db0-a807-664cf085391c/volumes" Dec 03 00:13:49 crc kubenswrapper[4953]: I1203 00:13:49.366639 4953 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-dpsq6"] Dec 03 00:13:49 crc kubenswrapper[4953]: E1203 00:13:49.366907 4953 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6b353eef-b7f9-4dbc-b2aa-9860220438cc" containerName="extract-utilities" Dec 03 00:13:49 crc kubenswrapper[4953]: I1203 00:13:49.366928 4953 state_mem.go:107] "Deleted CPUSet assignment" podUID="6b353eef-b7f9-4dbc-b2aa-9860220438cc" containerName="extract-utilities" Dec 03 00:13:49 crc kubenswrapper[4953]: E1203 00:13:49.366941 4953 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7124b81d-b87c-45de-a112-20602c4d9adb" containerName="marketplace-operator" Dec 03 00:13:49 crc kubenswrapper[4953]: I1203 00:13:49.366956 4953 state_mem.go:107] "Deleted CPUSet assignment" podUID="7124b81d-b87c-45de-a112-20602c4d9adb" containerName="marketplace-operator" Dec 03 00:13:49 crc kubenswrapper[4953]: E1203 00:13:49.366982 4953 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9064a5ab-24ee-4258-8363-6ce4ffa54b3f" containerName="extract-content" Dec 03 00:13:49 crc kubenswrapper[4953]: I1203 00:13:49.366989 4953 state_mem.go:107] "Deleted CPUSet assignment" podUID="9064a5ab-24ee-4258-8363-6ce4ffa54b3f" containerName="extract-content" Dec 03 00:13:49 crc kubenswrapper[4953]: E1203 00:13:49.367000 4953 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="52fd5773-3888-49f1-813d-10d262f4c901" containerName="registry-server" Dec 03 00:13:49 crc kubenswrapper[4953]: I1203 00:13:49.367007 4953 state_mem.go:107] "Deleted CPUSet assignment" podUID="52fd5773-3888-49f1-813d-10d262f4c901" containerName="registry-server" Dec 03 00:13:49 crc kubenswrapper[4953]: E1203 00:13:49.367017 4953 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="52fd5773-3888-49f1-813d-10d262f4c901" containerName="extract-content" Dec 03 00:13:49 crc kubenswrapper[4953]: I1203 00:13:49.367023 4953 state_mem.go:107] "Deleted CPUSet assignment" podUID="52fd5773-3888-49f1-813d-10d262f4c901" containerName="extract-content" Dec 03 00:13:49 crc kubenswrapper[4953]: E1203 00:13:49.367032 4953 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e98b8740-f6fc-4db0-a807-664cf085391c" containerName="registry-server" Dec 03 00:13:49 crc kubenswrapper[4953]: I1203 00:13:49.367039 4953 state_mem.go:107] "Deleted CPUSet assignment" podUID="e98b8740-f6fc-4db0-a807-664cf085391c" containerName="registry-server" Dec 03 00:13:49 crc kubenswrapper[4953]: E1203 00:13:49.367050 4953 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e98b8740-f6fc-4db0-a807-664cf085391c" containerName="extract-content" Dec 03 00:13:49 crc kubenswrapper[4953]: I1203 00:13:49.367056 4953 state_mem.go:107] "Deleted CPUSet assignment" podUID="e98b8740-f6fc-4db0-a807-664cf085391c" containerName="extract-content" Dec 03 00:13:49 crc kubenswrapper[4953]: E1203 00:13:49.367073 4953 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="52fd5773-3888-49f1-813d-10d262f4c901" containerName="extract-utilities" Dec 03 00:13:49 crc kubenswrapper[4953]: I1203 00:13:49.367079 4953 state_mem.go:107] "Deleted CPUSet assignment" podUID="52fd5773-3888-49f1-813d-10d262f4c901" containerName="extract-utilities" Dec 03 00:13:49 crc kubenswrapper[4953]: E1203 00:13:49.367087 4953 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6b353eef-b7f9-4dbc-b2aa-9860220438cc" containerName="extract-content" Dec 03 00:13:49 crc kubenswrapper[4953]: I1203 00:13:49.367094 4953 state_mem.go:107] "Deleted CPUSet assignment" podUID="6b353eef-b7f9-4dbc-b2aa-9860220438cc" containerName="extract-content" Dec 03 00:13:49 crc kubenswrapper[4953]: E1203 00:13:49.367103 4953 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6b353eef-b7f9-4dbc-b2aa-9860220438cc" containerName="registry-server" Dec 03 00:13:49 crc kubenswrapper[4953]: I1203 00:13:49.367111 4953 state_mem.go:107] "Deleted CPUSet assignment" podUID="6b353eef-b7f9-4dbc-b2aa-9860220438cc" containerName="registry-server" Dec 03 00:13:49 crc kubenswrapper[4953]: E1203 00:13:49.367127 4953 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e98b8740-f6fc-4db0-a807-664cf085391c" containerName="extract-utilities" Dec 03 00:13:49 crc kubenswrapper[4953]: I1203 00:13:49.367135 4953 state_mem.go:107] "Deleted CPUSet assignment" podUID="e98b8740-f6fc-4db0-a807-664cf085391c" containerName="extract-utilities" Dec 03 00:13:49 crc kubenswrapper[4953]: E1203 00:13:49.367147 4953 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9064a5ab-24ee-4258-8363-6ce4ffa54b3f" containerName="extract-utilities" Dec 03 00:13:49 crc kubenswrapper[4953]: I1203 00:13:49.367157 4953 state_mem.go:107] "Deleted CPUSet assignment" podUID="9064a5ab-24ee-4258-8363-6ce4ffa54b3f" containerName="extract-utilities" Dec 03 00:13:49 crc kubenswrapper[4953]: E1203 00:13:49.367170 4953 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9064a5ab-24ee-4258-8363-6ce4ffa54b3f" containerName="registry-server" Dec 03 00:13:49 crc kubenswrapper[4953]: I1203 00:13:49.367178 4953 state_mem.go:107] "Deleted CPUSet assignment" podUID="9064a5ab-24ee-4258-8363-6ce4ffa54b3f" containerName="registry-server" Dec 03 00:13:49 crc kubenswrapper[4953]: I1203 00:13:49.367290 4953 memory_manager.go:354] "RemoveStaleState removing state" podUID="e98b8740-f6fc-4db0-a807-664cf085391c" containerName="registry-server" Dec 03 00:13:49 crc kubenswrapper[4953]: I1203 00:13:49.367308 4953 memory_manager.go:354] "RemoveStaleState removing state" podUID="52fd5773-3888-49f1-813d-10d262f4c901" containerName="registry-server" Dec 03 00:13:49 crc kubenswrapper[4953]: I1203 00:13:49.367318 4953 memory_manager.go:354] "RemoveStaleState removing state" podUID="7124b81d-b87c-45de-a112-20602c4d9adb" containerName="marketplace-operator" Dec 03 00:13:49 crc kubenswrapper[4953]: I1203 00:13:49.367331 4953 memory_manager.go:354] "RemoveStaleState removing state" podUID="7124b81d-b87c-45de-a112-20602c4d9adb" containerName="marketplace-operator" Dec 03 00:13:49 crc kubenswrapper[4953]: I1203 00:13:49.367342 4953 memory_manager.go:354] "RemoveStaleState removing state" podUID="9064a5ab-24ee-4258-8363-6ce4ffa54b3f" containerName="registry-server" Dec 03 00:13:49 crc kubenswrapper[4953]: I1203 00:13:49.367353 4953 memory_manager.go:354] "RemoveStaleState removing state" podUID="6b353eef-b7f9-4dbc-b2aa-9860220438cc" containerName="registry-server" Dec 03 00:13:49 crc kubenswrapper[4953]: E1203 00:13:49.367475 4953 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7124b81d-b87c-45de-a112-20602c4d9adb" containerName="marketplace-operator" Dec 03 00:13:49 crc kubenswrapper[4953]: I1203 00:13:49.367486 4953 state_mem.go:107] "Deleted CPUSet assignment" podUID="7124b81d-b87c-45de-a112-20602c4d9adb" containerName="marketplace-operator" Dec 03 00:13:49 crc kubenswrapper[4953]: I1203 00:13:49.368465 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-dpsq6" Dec 03 00:13:49 crc kubenswrapper[4953]: I1203 00:13:49.372171 4953 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Dec 03 00:13:49 crc kubenswrapper[4953]: I1203 00:13:49.377745 4953 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-dpsq6"] Dec 03 00:13:49 crc kubenswrapper[4953]: I1203 00:13:49.544999 4953 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-lcrlw"] Dec 03 00:13:49 crc kubenswrapper[4953]: I1203 00:13:49.545078 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c2b69c71-8e67-461d-9ddf-cfb439ebb6e1-utilities\") pod \"redhat-marketplace-dpsq6\" (UID: \"c2b69c71-8e67-461d-9ddf-cfb439ebb6e1\") " pod="openshift-marketplace/redhat-marketplace-dpsq6" Dec 03 00:13:49 crc kubenswrapper[4953]: I1203 00:13:49.546445 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c2b69c71-8e67-461d-9ddf-cfb439ebb6e1-catalog-content\") pod \"redhat-marketplace-dpsq6\" (UID: \"c2b69c71-8e67-461d-9ddf-cfb439ebb6e1\") " pod="openshift-marketplace/redhat-marketplace-dpsq6" Dec 03 00:13:49 crc kubenswrapper[4953]: I1203 00:13:49.546532 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gb6nx\" (UniqueName: \"kubernetes.io/projected/c2b69c71-8e67-461d-9ddf-cfb439ebb6e1-kube-api-access-gb6nx\") pod \"redhat-marketplace-dpsq6\" (UID: \"c2b69c71-8e67-461d-9ddf-cfb439ebb6e1\") " pod="openshift-marketplace/redhat-marketplace-dpsq6" Dec 03 00:13:49 crc kubenswrapper[4953]: I1203 00:13:49.547524 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-lcrlw" Dec 03 00:13:49 crc kubenswrapper[4953]: I1203 00:13:49.550409 4953 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Dec 03 00:13:49 crc kubenswrapper[4953]: I1203 00:13:49.573910 4953 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-lcrlw"] Dec 03 00:13:49 crc kubenswrapper[4953]: I1203 00:13:49.651789 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/475ce63b-d94a-4b5f-b213-bbc2dd733e63-utilities\") pod \"certified-operators-lcrlw\" (UID: \"475ce63b-d94a-4b5f-b213-bbc2dd733e63\") " pod="openshift-marketplace/certified-operators-lcrlw" Dec 03 00:13:49 crc kubenswrapper[4953]: I1203 00:13:49.651867 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n4bd6\" (UniqueName: \"kubernetes.io/projected/475ce63b-d94a-4b5f-b213-bbc2dd733e63-kube-api-access-n4bd6\") pod \"certified-operators-lcrlw\" (UID: \"475ce63b-d94a-4b5f-b213-bbc2dd733e63\") " pod="openshift-marketplace/certified-operators-lcrlw" Dec 03 00:13:49 crc kubenswrapper[4953]: I1203 00:13:49.651902 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c2b69c71-8e67-461d-9ddf-cfb439ebb6e1-utilities\") pod \"redhat-marketplace-dpsq6\" (UID: \"c2b69c71-8e67-461d-9ddf-cfb439ebb6e1\") " pod="openshift-marketplace/redhat-marketplace-dpsq6" Dec 03 00:13:49 crc kubenswrapper[4953]: I1203 00:13:49.651943 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c2b69c71-8e67-461d-9ddf-cfb439ebb6e1-catalog-content\") pod \"redhat-marketplace-dpsq6\" (UID: \"c2b69c71-8e67-461d-9ddf-cfb439ebb6e1\") " pod="openshift-marketplace/redhat-marketplace-dpsq6" Dec 03 00:13:49 crc kubenswrapper[4953]: I1203 00:13:49.652007 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/475ce63b-d94a-4b5f-b213-bbc2dd733e63-catalog-content\") pod \"certified-operators-lcrlw\" (UID: \"475ce63b-d94a-4b5f-b213-bbc2dd733e63\") " pod="openshift-marketplace/certified-operators-lcrlw" Dec 03 00:13:49 crc kubenswrapper[4953]: I1203 00:13:49.652032 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gb6nx\" (UniqueName: \"kubernetes.io/projected/c2b69c71-8e67-461d-9ddf-cfb439ebb6e1-kube-api-access-gb6nx\") pod \"redhat-marketplace-dpsq6\" (UID: \"c2b69c71-8e67-461d-9ddf-cfb439ebb6e1\") " pod="openshift-marketplace/redhat-marketplace-dpsq6" Dec 03 00:13:49 crc kubenswrapper[4953]: I1203 00:13:49.652445 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c2b69c71-8e67-461d-9ddf-cfb439ebb6e1-utilities\") pod \"redhat-marketplace-dpsq6\" (UID: \"c2b69c71-8e67-461d-9ddf-cfb439ebb6e1\") " pod="openshift-marketplace/redhat-marketplace-dpsq6" Dec 03 00:13:49 crc kubenswrapper[4953]: I1203 00:13:49.652601 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c2b69c71-8e67-461d-9ddf-cfb439ebb6e1-catalog-content\") pod \"redhat-marketplace-dpsq6\" (UID: \"c2b69c71-8e67-461d-9ddf-cfb439ebb6e1\") " pod="openshift-marketplace/redhat-marketplace-dpsq6" Dec 03 00:13:49 crc kubenswrapper[4953]: I1203 00:13:49.674957 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gb6nx\" (UniqueName: \"kubernetes.io/projected/c2b69c71-8e67-461d-9ddf-cfb439ebb6e1-kube-api-access-gb6nx\") pod \"redhat-marketplace-dpsq6\" (UID: \"c2b69c71-8e67-461d-9ddf-cfb439ebb6e1\") " pod="openshift-marketplace/redhat-marketplace-dpsq6" Dec 03 00:13:49 crc kubenswrapper[4953]: I1203 00:13:49.691558 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-dpsq6" Dec 03 00:13:49 crc kubenswrapper[4953]: I1203 00:13:49.753947 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/475ce63b-d94a-4b5f-b213-bbc2dd733e63-utilities\") pod \"certified-operators-lcrlw\" (UID: \"475ce63b-d94a-4b5f-b213-bbc2dd733e63\") " pod="openshift-marketplace/certified-operators-lcrlw" Dec 03 00:13:49 crc kubenswrapper[4953]: I1203 00:13:49.754405 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n4bd6\" (UniqueName: \"kubernetes.io/projected/475ce63b-d94a-4b5f-b213-bbc2dd733e63-kube-api-access-n4bd6\") pod \"certified-operators-lcrlw\" (UID: \"475ce63b-d94a-4b5f-b213-bbc2dd733e63\") " pod="openshift-marketplace/certified-operators-lcrlw" Dec 03 00:13:49 crc kubenswrapper[4953]: I1203 00:13:49.754450 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/475ce63b-d94a-4b5f-b213-bbc2dd733e63-catalog-content\") pod \"certified-operators-lcrlw\" (UID: \"475ce63b-d94a-4b5f-b213-bbc2dd733e63\") " pod="openshift-marketplace/certified-operators-lcrlw" Dec 03 00:13:49 crc kubenswrapper[4953]: I1203 00:13:49.755070 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/475ce63b-d94a-4b5f-b213-bbc2dd733e63-utilities\") pod \"certified-operators-lcrlw\" (UID: \"475ce63b-d94a-4b5f-b213-bbc2dd733e63\") " pod="openshift-marketplace/certified-operators-lcrlw" Dec 03 00:13:49 crc kubenswrapper[4953]: I1203 00:13:49.755092 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/475ce63b-d94a-4b5f-b213-bbc2dd733e63-catalog-content\") pod \"certified-operators-lcrlw\" (UID: \"475ce63b-d94a-4b5f-b213-bbc2dd733e63\") " pod="openshift-marketplace/certified-operators-lcrlw" Dec 03 00:13:49 crc kubenswrapper[4953]: I1203 00:13:49.784480 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n4bd6\" (UniqueName: \"kubernetes.io/projected/475ce63b-d94a-4b5f-b213-bbc2dd733e63-kube-api-access-n4bd6\") pod \"certified-operators-lcrlw\" (UID: \"475ce63b-d94a-4b5f-b213-bbc2dd733e63\") " pod="openshift-marketplace/certified-operators-lcrlw" Dec 03 00:13:49 crc kubenswrapper[4953]: I1203 00:13:49.881441 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-lcrlw" Dec 03 00:13:51 crc kubenswrapper[4953]: I1203 00:13:50.040961 4953 generic.go:334] "Generic (PLEG): container finished" podID="53ffb1e9-9a2e-4869-80d7-776f4aa7ef8b" containerID="cd3fdc5de406f82703cc48138f69373c50d1862f3fb1f4cc4572aa863d2eca4c" exitCode=0 Dec 03 00:13:51 crc kubenswrapper[4953]: I1203 00:13:50.041030 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-6dfbn" event={"ID":"53ffb1e9-9a2e-4869-80d7-776f4aa7ef8b","Type":"ContainerDied","Data":"cd3fdc5de406f82703cc48138f69373c50d1862f3fb1f4cc4572aa863d2eca4c"} Dec 03 00:13:51 crc kubenswrapper[4953]: I1203 00:13:50.041121 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-6dfbn" event={"ID":"53ffb1e9-9a2e-4869-80d7-776f4aa7ef8b","Type":"ContainerStarted","Data":"257c3b050511c36fa41f2b2785b39713a95149ac43905452ee5690643ac49dc2"} Dec 03 00:13:51 crc kubenswrapper[4953]: I1203 00:13:50.041147 4953 scope.go:117] "RemoveContainer" containerID="7df84405eda520ad5a22d35750edf07ef3c3977a756ad5739c54f91cba9da7b2" Dec 03 00:13:51 crc kubenswrapper[4953]: I1203 00:13:51.397741 4953 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-dpsq6"] Dec 03 00:13:51 crc kubenswrapper[4953]: I1203 00:13:51.680114 4953 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-lcrlw"] Dec 03 00:13:51 crc kubenswrapper[4953]: W1203 00:13:51.694259 4953 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod475ce63b_d94a_4b5f_b213_bbc2dd733e63.slice/crio-9905ffd9296d716eaca22578368a795926e3e9c6e2e6e248cccb0fcafd2bc1e3 WatchSource:0}: Error finding container 9905ffd9296d716eaca22578368a795926e3e9c6e2e6e248cccb0fcafd2bc1e3: Status 404 returned error can't find the container with id 9905ffd9296d716eaca22578368a795926e3e9c6e2e6e248cccb0fcafd2bc1e3 Dec 03 00:13:51 crc kubenswrapper[4953]: I1203 00:13:51.755135 4953 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-lxqn5"] Dec 03 00:13:51 crc kubenswrapper[4953]: I1203 00:13:51.757439 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-lxqn5" Dec 03 00:13:51 crc kubenswrapper[4953]: I1203 00:13:51.763673 4953 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Dec 03 00:13:51 crc kubenswrapper[4953]: I1203 00:13:51.765456 4953 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-lxqn5"] Dec 03 00:13:51 crc kubenswrapper[4953]: I1203 00:13:51.808167 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/60145acc-39c4-483a-b635-f12937264ce0-catalog-content\") pod \"community-operators-lxqn5\" (UID: \"60145acc-39c4-483a-b635-f12937264ce0\") " pod="openshift-marketplace/community-operators-lxqn5" Dec 03 00:13:51 crc kubenswrapper[4953]: I1203 00:13:51.808296 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/60145acc-39c4-483a-b635-f12937264ce0-utilities\") pod \"community-operators-lxqn5\" (UID: \"60145acc-39c4-483a-b635-f12937264ce0\") " pod="openshift-marketplace/community-operators-lxqn5" Dec 03 00:13:51 crc kubenswrapper[4953]: I1203 00:13:51.808352 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4dpk7\" (UniqueName: \"kubernetes.io/projected/60145acc-39c4-483a-b635-f12937264ce0-kube-api-access-4dpk7\") pod \"community-operators-lxqn5\" (UID: \"60145acc-39c4-483a-b635-f12937264ce0\") " pod="openshift-marketplace/community-operators-lxqn5" Dec 03 00:13:51 crc kubenswrapper[4953]: I1203 00:13:51.909806 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4dpk7\" (UniqueName: \"kubernetes.io/projected/60145acc-39c4-483a-b635-f12937264ce0-kube-api-access-4dpk7\") pod \"community-operators-lxqn5\" (UID: \"60145acc-39c4-483a-b635-f12937264ce0\") " pod="openshift-marketplace/community-operators-lxqn5" Dec 03 00:13:51 crc kubenswrapper[4953]: I1203 00:13:51.909907 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/60145acc-39c4-483a-b635-f12937264ce0-catalog-content\") pod \"community-operators-lxqn5\" (UID: \"60145acc-39c4-483a-b635-f12937264ce0\") " pod="openshift-marketplace/community-operators-lxqn5" Dec 03 00:13:51 crc kubenswrapper[4953]: I1203 00:13:51.909961 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/60145acc-39c4-483a-b635-f12937264ce0-utilities\") pod \"community-operators-lxqn5\" (UID: \"60145acc-39c4-483a-b635-f12937264ce0\") " pod="openshift-marketplace/community-operators-lxqn5" Dec 03 00:13:51 crc kubenswrapper[4953]: I1203 00:13:51.910997 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/60145acc-39c4-483a-b635-f12937264ce0-utilities\") pod \"community-operators-lxqn5\" (UID: \"60145acc-39c4-483a-b635-f12937264ce0\") " pod="openshift-marketplace/community-operators-lxqn5" Dec 03 00:13:51 crc kubenswrapper[4953]: I1203 00:13:51.913497 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/60145acc-39c4-483a-b635-f12937264ce0-catalog-content\") pod \"community-operators-lxqn5\" (UID: \"60145acc-39c4-483a-b635-f12937264ce0\") " pod="openshift-marketplace/community-operators-lxqn5" Dec 03 00:13:51 crc kubenswrapper[4953]: I1203 00:13:51.940155 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4dpk7\" (UniqueName: \"kubernetes.io/projected/60145acc-39c4-483a-b635-f12937264ce0-kube-api-access-4dpk7\") pod \"community-operators-lxqn5\" (UID: \"60145acc-39c4-483a-b635-f12937264ce0\") " pod="openshift-marketplace/community-operators-lxqn5" Dec 03 00:13:51 crc kubenswrapper[4953]: I1203 00:13:51.956497 4953 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-qrq2c"] Dec 03 00:13:51 crc kubenswrapper[4953]: I1203 00:13:51.958043 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-qrq2c" Dec 03 00:13:51 crc kubenswrapper[4953]: I1203 00:13:51.961805 4953 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Dec 03 00:13:51 crc kubenswrapper[4953]: I1203 00:13:51.963084 4953 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-qrq2c"] Dec 03 00:13:52 crc kubenswrapper[4953]: I1203 00:13:52.011294 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9b3f5bab-705f-4fa1-8f74-6ae6a6aac120-utilities\") pod \"redhat-operators-qrq2c\" (UID: \"9b3f5bab-705f-4fa1-8f74-6ae6a6aac120\") " pod="openshift-marketplace/redhat-operators-qrq2c" Dec 03 00:13:52 crc kubenswrapper[4953]: I1203 00:13:52.011349 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9b3f5bab-705f-4fa1-8f74-6ae6a6aac120-catalog-content\") pod \"redhat-operators-qrq2c\" (UID: \"9b3f5bab-705f-4fa1-8f74-6ae6a6aac120\") " pod="openshift-marketplace/redhat-operators-qrq2c" Dec 03 00:13:52 crc kubenswrapper[4953]: I1203 00:13:52.011384 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6rt2l\" (UniqueName: \"kubernetes.io/projected/9b3f5bab-705f-4fa1-8f74-6ae6a6aac120-kube-api-access-6rt2l\") pod \"redhat-operators-qrq2c\" (UID: \"9b3f5bab-705f-4fa1-8f74-6ae6a6aac120\") " pod="openshift-marketplace/redhat-operators-qrq2c" Dec 03 00:13:52 crc kubenswrapper[4953]: I1203 00:13:52.112595 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9b3f5bab-705f-4fa1-8f74-6ae6a6aac120-utilities\") pod \"redhat-operators-qrq2c\" (UID: \"9b3f5bab-705f-4fa1-8f74-6ae6a6aac120\") " pod="openshift-marketplace/redhat-operators-qrq2c" Dec 03 00:13:52 crc kubenswrapper[4953]: I1203 00:13:52.112668 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9b3f5bab-705f-4fa1-8f74-6ae6a6aac120-catalog-content\") pod \"redhat-operators-qrq2c\" (UID: \"9b3f5bab-705f-4fa1-8f74-6ae6a6aac120\") " pod="openshift-marketplace/redhat-operators-qrq2c" Dec 03 00:13:52 crc kubenswrapper[4953]: I1203 00:13:52.112711 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6rt2l\" (UniqueName: \"kubernetes.io/projected/9b3f5bab-705f-4fa1-8f74-6ae6a6aac120-kube-api-access-6rt2l\") pod \"redhat-operators-qrq2c\" (UID: \"9b3f5bab-705f-4fa1-8f74-6ae6a6aac120\") " pod="openshift-marketplace/redhat-operators-qrq2c" Dec 03 00:13:52 crc kubenswrapper[4953]: I1203 00:13:52.113405 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9b3f5bab-705f-4fa1-8f74-6ae6a6aac120-utilities\") pod \"redhat-operators-qrq2c\" (UID: \"9b3f5bab-705f-4fa1-8f74-6ae6a6aac120\") " pod="openshift-marketplace/redhat-operators-qrq2c" Dec 03 00:13:52 crc kubenswrapper[4953]: I1203 00:13:52.113504 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9b3f5bab-705f-4fa1-8f74-6ae6a6aac120-catalog-content\") pod \"redhat-operators-qrq2c\" (UID: \"9b3f5bab-705f-4fa1-8f74-6ae6a6aac120\") " pod="openshift-marketplace/redhat-operators-qrq2c" Dec 03 00:13:52 crc kubenswrapper[4953]: I1203 00:13:52.127881 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-lxqn5" Dec 03 00:13:52 crc kubenswrapper[4953]: I1203 00:13:52.130229 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6rt2l\" (UniqueName: \"kubernetes.io/projected/9b3f5bab-705f-4fa1-8f74-6ae6a6aac120-kube-api-access-6rt2l\") pod \"redhat-operators-qrq2c\" (UID: \"9b3f5bab-705f-4fa1-8f74-6ae6a6aac120\") " pod="openshift-marketplace/redhat-operators-qrq2c" Dec 03 00:13:52 crc kubenswrapper[4953]: I1203 00:13:52.315783 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-qrq2c" Dec 03 00:13:52 crc kubenswrapper[4953]: I1203 00:13:52.331921 4953 generic.go:334] "Generic (PLEG): container finished" podID="475ce63b-d94a-4b5f-b213-bbc2dd733e63" containerID="895b09bae68edbcd83b2bc328a54568dd1da51e8ce999dae6b062b2ad3813e80" exitCode=0 Dec 03 00:13:52 crc kubenswrapper[4953]: I1203 00:13:52.332044 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lcrlw" event={"ID":"475ce63b-d94a-4b5f-b213-bbc2dd733e63","Type":"ContainerDied","Data":"895b09bae68edbcd83b2bc328a54568dd1da51e8ce999dae6b062b2ad3813e80"} Dec 03 00:13:52 crc kubenswrapper[4953]: I1203 00:13:52.332278 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lcrlw" event={"ID":"475ce63b-d94a-4b5f-b213-bbc2dd733e63","Type":"ContainerStarted","Data":"9905ffd9296d716eaca22578368a795926e3e9c6e2e6e248cccb0fcafd2bc1e3"} Dec 03 00:13:52 crc kubenswrapper[4953]: I1203 00:13:52.354409 4953 generic.go:334] "Generic (PLEG): container finished" podID="c2b69c71-8e67-461d-9ddf-cfb439ebb6e1" containerID="b915762a27acbf96f1aad344c6de71c18cb55d5309f14ee44012759acda59245" exitCode=0 Dec 03 00:13:52 crc kubenswrapper[4953]: I1203 00:13:52.354480 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dpsq6" event={"ID":"c2b69c71-8e67-461d-9ddf-cfb439ebb6e1","Type":"ContainerDied","Data":"b915762a27acbf96f1aad344c6de71c18cb55d5309f14ee44012759acda59245"} Dec 03 00:13:52 crc kubenswrapper[4953]: I1203 00:13:52.354502 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dpsq6" event={"ID":"c2b69c71-8e67-461d-9ddf-cfb439ebb6e1","Type":"ContainerStarted","Data":"199c1982390394e5cde47a5fbe1309d30350e3d7d92ac8ca558ca49c7b125c0b"} Dec 03 00:13:52 crc kubenswrapper[4953]: I1203 00:13:52.605227 4953 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-lxqn5"] Dec 03 00:13:52 crc kubenswrapper[4953]: W1203 00:13:52.612153 4953 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod60145acc_39c4_483a_b635_f12937264ce0.slice/crio-b347e2826a6a5344ac6b0acdf946a827c6554dcd37dc7ab2afa37b4c9d235a0c WatchSource:0}: Error finding container b347e2826a6a5344ac6b0acdf946a827c6554dcd37dc7ab2afa37b4c9d235a0c: Status 404 returned error can't find the container with id b347e2826a6a5344ac6b0acdf946a827c6554dcd37dc7ab2afa37b4c9d235a0c Dec 03 00:13:52 crc kubenswrapper[4953]: I1203 00:13:52.773524 4953 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-qrq2c"] Dec 03 00:13:52 crc kubenswrapper[4953]: W1203 00:13:52.781324 4953 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9b3f5bab_705f_4fa1_8f74_6ae6a6aac120.slice/crio-a5270c5b65a786f1061e2a6e004550339ae7c7e2d15f3fad733d404837d0bfe7 WatchSource:0}: Error finding container a5270c5b65a786f1061e2a6e004550339ae7c7e2d15f3fad733d404837d0bfe7: Status 404 returned error can't find the container with id a5270c5b65a786f1061e2a6e004550339ae7c7e2d15f3fad733d404837d0bfe7 Dec 03 00:13:53 crc kubenswrapper[4953]: I1203 00:13:53.370480 4953 generic.go:334] "Generic (PLEG): container finished" podID="60145acc-39c4-483a-b635-f12937264ce0" containerID="12bf0bc51b906ec0b225bb613ec1eeceda2706bbae04ad7fbc195fc33441c120" exitCode=0 Dec 03 00:13:53 crc kubenswrapper[4953]: I1203 00:13:53.370624 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lxqn5" event={"ID":"60145acc-39c4-483a-b635-f12937264ce0","Type":"ContainerDied","Data":"12bf0bc51b906ec0b225bb613ec1eeceda2706bbae04ad7fbc195fc33441c120"} Dec 03 00:13:53 crc kubenswrapper[4953]: I1203 00:13:53.370659 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lxqn5" event={"ID":"60145acc-39c4-483a-b635-f12937264ce0","Type":"ContainerStarted","Data":"b347e2826a6a5344ac6b0acdf946a827c6554dcd37dc7ab2afa37b4c9d235a0c"} Dec 03 00:13:53 crc kubenswrapper[4953]: I1203 00:13:53.373555 4953 generic.go:334] "Generic (PLEG): container finished" podID="9b3f5bab-705f-4fa1-8f74-6ae6a6aac120" containerID="1f53ee2f167b81458827e86194ac0a6276d0b47b78f1ec99a510fc88fc6425b3" exitCode=0 Dec 03 00:13:53 crc kubenswrapper[4953]: I1203 00:13:53.373603 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qrq2c" event={"ID":"9b3f5bab-705f-4fa1-8f74-6ae6a6aac120","Type":"ContainerDied","Data":"1f53ee2f167b81458827e86194ac0a6276d0b47b78f1ec99a510fc88fc6425b3"} Dec 03 00:13:53 crc kubenswrapper[4953]: I1203 00:13:53.373633 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qrq2c" event={"ID":"9b3f5bab-705f-4fa1-8f74-6ae6a6aac120","Type":"ContainerStarted","Data":"a5270c5b65a786f1061e2a6e004550339ae7c7e2d15f3fad733d404837d0bfe7"} Dec 03 00:13:54 crc kubenswrapper[4953]: I1203 00:13:54.386394 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qrq2c" event={"ID":"9b3f5bab-705f-4fa1-8f74-6ae6a6aac120","Type":"ContainerStarted","Data":"8a7e846c367cbe5195abd3bc78215e50199c5a18393509081176bf4b71d46c7e"} Dec 03 00:13:54 crc kubenswrapper[4953]: I1203 00:13:54.396111 4953 generic.go:334] "Generic (PLEG): container finished" podID="c2b69c71-8e67-461d-9ddf-cfb439ebb6e1" containerID="bc8121247f710ab2d28dcd715750c2c07d278ab469c3c0361138ab85f78b76ca" exitCode=0 Dec 03 00:13:54 crc kubenswrapper[4953]: I1203 00:13:54.396373 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dpsq6" event={"ID":"c2b69c71-8e67-461d-9ddf-cfb439ebb6e1","Type":"ContainerDied","Data":"bc8121247f710ab2d28dcd715750c2c07d278ab469c3c0361138ab85f78b76ca"} Dec 03 00:13:54 crc kubenswrapper[4953]: I1203 00:13:54.405264 4953 generic.go:334] "Generic (PLEG): container finished" podID="475ce63b-d94a-4b5f-b213-bbc2dd733e63" containerID="f43df30325746a0ea651bda7463e385fa0a5b80e150df8a2bb1ad35ccac30822" exitCode=0 Dec 03 00:13:54 crc kubenswrapper[4953]: I1203 00:13:54.405383 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lcrlw" event={"ID":"475ce63b-d94a-4b5f-b213-bbc2dd733e63","Type":"ContainerDied","Data":"f43df30325746a0ea651bda7463e385fa0a5b80e150df8a2bb1ad35ccac30822"} Dec 03 00:13:54 crc kubenswrapper[4953]: I1203 00:13:54.417840 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lxqn5" event={"ID":"60145acc-39c4-483a-b635-f12937264ce0","Type":"ContainerStarted","Data":"d17f906138ac1b7e8590039cd7d8f5b98664d369465b10d2d14959d0ac406b11"} Dec 03 00:13:55 crc kubenswrapper[4953]: I1203 00:13:55.428090 4953 generic.go:334] "Generic (PLEG): container finished" podID="9b3f5bab-705f-4fa1-8f74-6ae6a6aac120" containerID="8a7e846c367cbe5195abd3bc78215e50199c5a18393509081176bf4b71d46c7e" exitCode=0 Dec 03 00:13:55 crc kubenswrapper[4953]: I1203 00:13:55.428154 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qrq2c" event={"ID":"9b3f5bab-705f-4fa1-8f74-6ae6a6aac120","Type":"ContainerDied","Data":"8a7e846c367cbe5195abd3bc78215e50199c5a18393509081176bf4b71d46c7e"} Dec 03 00:13:55 crc kubenswrapper[4953]: I1203 00:13:55.433919 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dpsq6" event={"ID":"c2b69c71-8e67-461d-9ddf-cfb439ebb6e1","Type":"ContainerStarted","Data":"034c2b2807e39fd7848b03f21692b9a4094c04152d8464a2ca8f5d82217eb39a"} Dec 03 00:13:55 crc kubenswrapper[4953]: I1203 00:13:55.446399 4953 generic.go:334] "Generic (PLEG): container finished" podID="60145acc-39c4-483a-b635-f12937264ce0" containerID="d17f906138ac1b7e8590039cd7d8f5b98664d369465b10d2d14959d0ac406b11" exitCode=0 Dec 03 00:13:55 crc kubenswrapper[4953]: I1203 00:13:55.446445 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lxqn5" event={"ID":"60145acc-39c4-483a-b635-f12937264ce0","Type":"ContainerDied","Data":"d17f906138ac1b7e8590039cd7d8f5b98664d369465b10d2d14959d0ac406b11"} Dec 03 00:13:55 crc kubenswrapper[4953]: I1203 00:13:55.482260 4953 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-dpsq6" podStartSLOduration=3.975381496 podStartE2EDuration="6.482239246s" podCreationTimestamp="2025-12-03 00:13:49 +0000 UTC" firstStartedPulling="2025-12-03 00:13:52.361610514 +0000 UTC m=+421.345891297" lastFinishedPulling="2025-12-03 00:13:54.868468254 +0000 UTC m=+423.852749047" observedRunningTime="2025-12-03 00:13:55.479869796 +0000 UTC m=+424.464150589" watchObservedRunningTime="2025-12-03 00:13:55.482239246 +0000 UTC m=+424.466520049" Dec 03 00:13:55 crc kubenswrapper[4953]: I1203 00:13:55.498115 4953 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7447957dcb-m2cwp"] Dec 03 00:13:55 crc kubenswrapper[4953]: I1203 00:13:55.498588 4953 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-7447957dcb-m2cwp" podUID="fe523418-5f1d-472a-bf22-246672d50480" containerName="route-controller-manager" containerID="cri-o://87d1955f38cf491a4e593c5b87b7335bc3ed7e5bbe7ddfe055fe99731e169aac" gracePeriod=30 Dec 03 00:13:56 crc kubenswrapper[4953]: I1203 00:13:56.088398 4953 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-7447957dcb-m2cwp" Dec 03 00:13:56 crc kubenswrapper[4953]: I1203 00:13:56.188768 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qfrlj\" (UniqueName: \"kubernetes.io/projected/fe523418-5f1d-472a-bf22-246672d50480-kube-api-access-qfrlj\") pod \"fe523418-5f1d-472a-bf22-246672d50480\" (UID: \"fe523418-5f1d-472a-bf22-246672d50480\") " Dec 03 00:13:56 crc kubenswrapper[4953]: I1203 00:13:56.188812 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fe523418-5f1d-472a-bf22-246672d50480-config\") pod \"fe523418-5f1d-472a-bf22-246672d50480\" (UID: \"fe523418-5f1d-472a-bf22-246672d50480\") " Dec 03 00:13:56 crc kubenswrapper[4953]: I1203 00:13:56.188868 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/fe523418-5f1d-472a-bf22-246672d50480-client-ca\") pod \"fe523418-5f1d-472a-bf22-246672d50480\" (UID: \"fe523418-5f1d-472a-bf22-246672d50480\") " Dec 03 00:13:56 crc kubenswrapper[4953]: I1203 00:13:56.188910 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fe523418-5f1d-472a-bf22-246672d50480-serving-cert\") pod \"fe523418-5f1d-472a-bf22-246672d50480\" (UID: \"fe523418-5f1d-472a-bf22-246672d50480\") " Dec 03 00:13:56 crc kubenswrapper[4953]: I1203 00:13:56.190588 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fe523418-5f1d-472a-bf22-246672d50480-config" (OuterVolumeSpecName: "config") pod "fe523418-5f1d-472a-bf22-246672d50480" (UID: "fe523418-5f1d-472a-bf22-246672d50480"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:13:56 crc kubenswrapper[4953]: I1203 00:13:56.190605 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fe523418-5f1d-472a-bf22-246672d50480-client-ca" (OuterVolumeSpecName: "client-ca") pod "fe523418-5f1d-472a-bf22-246672d50480" (UID: "fe523418-5f1d-472a-bf22-246672d50480"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:13:56 crc kubenswrapper[4953]: I1203 00:13:56.201310 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fe523418-5f1d-472a-bf22-246672d50480-kube-api-access-qfrlj" (OuterVolumeSpecName: "kube-api-access-qfrlj") pod "fe523418-5f1d-472a-bf22-246672d50480" (UID: "fe523418-5f1d-472a-bf22-246672d50480"). InnerVolumeSpecName "kube-api-access-qfrlj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:13:56 crc kubenswrapper[4953]: I1203 00:13:56.201406 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fe523418-5f1d-472a-bf22-246672d50480-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "fe523418-5f1d-472a-bf22-246672d50480" (UID: "fe523418-5f1d-472a-bf22-246672d50480"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:13:56 crc kubenswrapper[4953]: I1203 00:13:56.290932 4953 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qfrlj\" (UniqueName: \"kubernetes.io/projected/fe523418-5f1d-472a-bf22-246672d50480-kube-api-access-qfrlj\") on node \"crc\" DevicePath \"\"" Dec 03 00:13:56 crc kubenswrapper[4953]: I1203 00:13:56.291498 4953 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fe523418-5f1d-472a-bf22-246672d50480-config\") on node \"crc\" DevicePath \"\"" Dec 03 00:13:56 crc kubenswrapper[4953]: I1203 00:13:56.291511 4953 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/fe523418-5f1d-472a-bf22-246672d50480-client-ca\") on node \"crc\" DevicePath \"\"" Dec 03 00:13:56 crc kubenswrapper[4953]: I1203 00:13:56.291522 4953 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fe523418-5f1d-472a-bf22-246672d50480-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 00:13:56 crc kubenswrapper[4953]: I1203 00:13:56.452921 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qrq2c" event={"ID":"9b3f5bab-705f-4fa1-8f74-6ae6a6aac120","Type":"ContainerStarted","Data":"f8d2a448152ded366148daf8aa9d5994cd85d94ed83b8e8e96643935401dd67a"} Dec 03 00:13:56 crc kubenswrapper[4953]: I1203 00:13:56.454753 4953 generic.go:334] "Generic (PLEG): container finished" podID="fe523418-5f1d-472a-bf22-246672d50480" containerID="87d1955f38cf491a4e593c5b87b7335bc3ed7e5bbe7ddfe055fe99731e169aac" exitCode=0 Dec 03 00:13:56 crc kubenswrapper[4953]: I1203 00:13:56.454837 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-7447957dcb-m2cwp" event={"ID":"fe523418-5f1d-472a-bf22-246672d50480","Type":"ContainerDied","Data":"87d1955f38cf491a4e593c5b87b7335bc3ed7e5bbe7ddfe055fe99731e169aac"} Dec 03 00:13:56 crc kubenswrapper[4953]: I1203 00:13:56.454861 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-7447957dcb-m2cwp" event={"ID":"fe523418-5f1d-472a-bf22-246672d50480","Type":"ContainerDied","Data":"796b1530b17e4b7ecf299cb7af0492a118cafda3857b9b32eb3a73056244e8c9"} Dec 03 00:13:56 crc kubenswrapper[4953]: I1203 00:13:56.454883 4953 scope.go:117] "RemoveContainer" containerID="87d1955f38cf491a4e593c5b87b7335bc3ed7e5bbe7ddfe055fe99731e169aac" Dec 03 00:13:56 crc kubenswrapper[4953]: I1203 00:13:56.455083 4953 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-7447957dcb-m2cwp" Dec 03 00:13:56 crc kubenswrapper[4953]: I1203 00:13:56.458710 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lcrlw" event={"ID":"475ce63b-d94a-4b5f-b213-bbc2dd733e63","Type":"ContainerStarted","Data":"b0ac7ef82cf8f3f4795828e4a7061147e9cad572e41432850216c81adf9aef17"} Dec 03 00:13:56 crc kubenswrapper[4953]: I1203 00:13:56.467202 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lxqn5" event={"ID":"60145acc-39c4-483a-b635-f12937264ce0","Type":"ContainerStarted","Data":"24d0d919e4aae7311b63d26fa6276ade68e116478d9d931a5fb0f3f29c1b54b1"} Dec 03 00:13:56 crc kubenswrapper[4953]: I1203 00:13:56.476461 4953 scope.go:117] "RemoveContainer" containerID="87d1955f38cf491a4e593c5b87b7335bc3ed7e5bbe7ddfe055fe99731e169aac" Dec 03 00:13:56 crc kubenswrapper[4953]: E1203 00:13:56.477101 4953 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"87d1955f38cf491a4e593c5b87b7335bc3ed7e5bbe7ddfe055fe99731e169aac\": container with ID starting with 87d1955f38cf491a4e593c5b87b7335bc3ed7e5bbe7ddfe055fe99731e169aac not found: ID does not exist" containerID="87d1955f38cf491a4e593c5b87b7335bc3ed7e5bbe7ddfe055fe99731e169aac" Dec 03 00:13:56 crc kubenswrapper[4953]: I1203 00:13:56.477156 4953 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"87d1955f38cf491a4e593c5b87b7335bc3ed7e5bbe7ddfe055fe99731e169aac"} err="failed to get container status \"87d1955f38cf491a4e593c5b87b7335bc3ed7e5bbe7ddfe055fe99731e169aac\": rpc error: code = NotFound desc = could not find container \"87d1955f38cf491a4e593c5b87b7335bc3ed7e5bbe7ddfe055fe99731e169aac\": container with ID starting with 87d1955f38cf491a4e593c5b87b7335bc3ed7e5bbe7ddfe055fe99731e169aac not found: ID does not exist" Dec 03 00:13:56 crc kubenswrapper[4953]: I1203 00:13:56.491237 4953 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-qrq2c" podStartSLOduration=2.822768466 podStartE2EDuration="5.491206481s" podCreationTimestamp="2025-12-03 00:13:51 +0000 UTC" firstStartedPulling="2025-12-03 00:13:53.377569235 +0000 UTC m=+422.361850018" lastFinishedPulling="2025-12-03 00:13:56.04600725 +0000 UTC m=+425.030288033" observedRunningTime="2025-12-03 00:13:56.486069272 +0000 UTC m=+425.470350065" watchObservedRunningTime="2025-12-03 00:13:56.491206481 +0000 UTC m=+425.475487264" Dec 03 00:13:56 crc kubenswrapper[4953]: I1203 00:13:56.538406 4953 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-lcrlw" podStartSLOduration=4.568064918 podStartE2EDuration="7.538377948s" podCreationTimestamp="2025-12-03 00:13:49 +0000 UTC" firstStartedPulling="2025-12-03 00:13:52.334137973 +0000 UTC m=+421.318418756" lastFinishedPulling="2025-12-03 00:13:55.304451003 +0000 UTC m=+424.288731786" observedRunningTime="2025-12-03 00:13:56.519915903 +0000 UTC m=+425.504196696" watchObservedRunningTime="2025-12-03 00:13:56.538377948 +0000 UTC m=+425.522658731" Dec 03 00:13:56 crc kubenswrapper[4953]: I1203 00:13:56.553847 4953 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-lxqn5" podStartSLOduration=2.9898307490000002 podStartE2EDuration="5.553825626s" podCreationTimestamp="2025-12-03 00:13:51 +0000 UTC" firstStartedPulling="2025-12-03 00:13:53.376904918 +0000 UTC m=+422.361185701" lastFinishedPulling="2025-12-03 00:13:55.940899795 +0000 UTC m=+424.925180578" observedRunningTime="2025-12-03 00:13:56.551527148 +0000 UTC m=+425.535807931" watchObservedRunningTime="2025-12-03 00:13:56.553825626 +0000 UTC m=+425.538106409" Dec 03 00:13:56 crc kubenswrapper[4953]: I1203 00:13:56.574257 4953 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7447957dcb-m2cwp"] Dec 03 00:13:56 crc kubenswrapper[4953]: I1203 00:13:56.580636 4953 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7447957dcb-m2cwp"] Dec 03 00:13:57 crc kubenswrapper[4953]: I1203 00:13:57.170990 4953 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6c9bb59cb6-hxtzb"] Dec 03 00:13:57 crc kubenswrapper[4953]: E1203 00:13:57.171878 4953 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fe523418-5f1d-472a-bf22-246672d50480" containerName="route-controller-manager" Dec 03 00:13:57 crc kubenswrapper[4953]: I1203 00:13:57.171900 4953 state_mem.go:107] "Deleted CPUSet assignment" podUID="fe523418-5f1d-472a-bf22-246672d50480" containerName="route-controller-manager" Dec 03 00:13:57 crc kubenswrapper[4953]: I1203 00:13:57.172091 4953 memory_manager.go:354] "RemoveStaleState removing state" podUID="fe523418-5f1d-472a-bf22-246672d50480" containerName="route-controller-manager" Dec 03 00:13:57 crc kubenswrapper[4953]: I1203 00:13:57.172753 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6c9bb59cb6-hxtzb" Dec 03 00:13:57 crc kubenswrapper[4953]: I1203 00:13:57.179372 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Dec 03 00:13:57 crc kubenswrapper[4953]: I1203 00:13:57.179632 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Dec 03 00:13:57 crc kubenswrapper[4953]: I1203 00:13:57.181184 4953 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6c9bb59cb6-hxtzb"] Dec 03 00:13:57 crc kubenswrapper[4953]: I1203 00:13:57.181219 4953 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Dec 03 00:13:57 crc kubenswrapper[4953]: I1203 00:13:57.181316 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Dec 03 00:13:57 crc kubenswrapper[4953]: I1203 00:13:57.181459 4953 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Dec 03 00:13:57 crc kubenswrapper[4953]: I1203 00:13:57.183395 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Dec 03 00:13:57 crc kubenswrapper[4953]: I1203 00:13:57.235387 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l6vk7\" (UniqueName: \"kubernetes.io/projected/0b5291af-1bf9-43e2-a128-a421d69c67ea-kube-api-access-l6vk7\") pod \"route-controller-manager-6c9bb59cb6-hxtzb\" (UID: \"0b5291af-1bf9-43e2-a128-a421d69c67ea\") " pod="openshift-route-controller-manager/route-controller-manager-6c9bb59cb6-hxtzb" Dec 03 00:13:57 crc kubenswrapper[4953]: I1203 00:13:57.235491 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0b5291af-1bf9-43e2-a128-a421d69c67ea-config\") pod \"route-controller-manager-6c9bb59cb6-hxtzb\" (UID: \"0b5291af-1bf9-43e2-a128-a421d69c67ea\") " pod="openshift-route-controller-manager/route-controller-manager-6c9bb59cb6-hxtzb" Dec 03 00:13:57 crc kubenswrapper[4953]: I1203 00:13:57.235525 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b5291af-1bf9-43e2-a128-a421d69c67ea-serving-cert\") pod \"route-controller-manager-6c9bb59cb6-hxtzb\" (UID: \"0b5291af-1bf9-43e2-a128-a421d69c67ea\") " pod="openshift-route-controller-manager/route-controller-manager-6c9bb59cb6-hxtzb" Dec 03 00:13:57 crc kubenswrapper[4953]: I1203 00:13:57.235588 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/0b5291af-1bf9-43e2-a128-a421d69c67ea-client-ca\") pod \"route-controller-manager-6c9bb59cb6-hxtzb\" (UID: \"0b5291af-1bf9-43e2-a128-a421d69c67ea\") " pod="openshift-route-controller-manager/route-controller-manager-6c9bb59cb6-hxtzb" Dec 03 00:13:57 crc kubenswrapper[4953]: I1203 00:13:57.336642 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l6vk7\" (UniqueName: \"kubernetes.io/projected/0b5291af-1bf9-43e2-a128-a421d69c67ea-kube-api-access-l6vk7\") pod \"route-controller-manager-6c9bb59cb6-hxtzb\" (UID: \"0b5291af-1bf9-43e2-a128-a421d69c67ea\") " pod="openshift-route-controller-manager/route-controller-manager-6c9bb59cb6-hxtzb" Dec 03 00:13:57 crc kubenswrapper[4953]: I1203 00:13:57.336723 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0b5291af-1bf9-43e2-a128-a421d69c67ea-config\") pod \"route-controller-manager-6c9bb59cb6-hxtzb\" (UID: \"0b5291af-1bf9-43e2-a128-a421d69c67ea\") " pod="openshift-route-controller-manager/route-controller-manager-6c9bb59cb6-hxtzb" Dec 03 00:13:57 crc kubenswrapper[4953]: I1203 00:13:57.336759 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b5291af-1bf9-43e2-a128-a421d69c67ea-serving-cert\") pod \"route-controller-manager-6c9bb59cb6-hxtzb\" (UID: \"0b5291af-1bf9-43e2-a128-a421d69c67ea\") " pod="openshift-route-controller-manager/route-controller-manager-6c9bb59cb6-hxtzb" Dec 03 00:13:57 crc kubenswrapper[4953]: I1203 00:13:57.336821 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/0b5291af-1bf9-43e2-a128-a421d69c67ea-client-ca\") pod \"route-controller-manager-6c9bb59cb6-hxtzb\" (UID: \"0b5291af-1bf9-43e2-a128-a421d69c67ea\") " pod="openshift-route-controller-manager/route-controller-manager-6c9bb59cb6-hxtzb" Dec 03 00:13:57 crc kubenswrapper[4953]: I1203 00:13:57.338150 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/0b5291af-1bf9-43e2-a128-a421d69c67ea-client-ca\") pod \"route-controller-manager-6c9bb59cb6-hxtzb\" (UID: \"0b5291af-1bf9-43e2-a128-a421d69c67ea\") " pod="openshift-route-controller-manager/route-controller-manager-6c9bb59cb6-hxtzb" Dec 03 00:13:57 crc kubenswrapper[4953]: I1203 00:13:57.339630 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0b5291af-1bf9-43e2-a128-a421d69c67ea-config\") pod \"route-controller-manager-6c9bb59cb6-hxtzb\" (UID: \"0b5291af-1bf9-43e2-a128-a421d69c67ea\") " pod="openshift-route-controller-manager/route-controller-manager-6c9bb59cb6-hxtzb" Dec 03 00:13:57 crc kubenswrapper[4953]: I1203 00:13:57.344281 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b5291af-1bf9-43e2-a128-a421d69c67ea-serving-cert\") pod \"route-controller-manager-6c9bb59cb6-hxtzb\" (UID: \"0b5291af-1bf9-43e2-a128-a421d69c67ea\") " pod="openshift-route-controller-manager/route-controller-manager-6c9bb59cb6-hxtzb" Dec 03 00:13:57 crc kubenswrapper[4953]: I1203 00:13:57.354605 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l6vk7\" (UniqueName: \"kubernetes.io/projected/0b5291af-1bf9-43e2-a128-a421d69c67ea-kube-api-access-l6vk7\") pod \"route-controller-manager-6c9bb59cb6-hxtzb\" (UID: \"0b5291af-1bf9-43e2-a128-a421d69c67ea\") " pod="openshift-route-controller-manager/route-controller-manager-6c9bb59cb6-hxtzb" Dec 03 00:13:57 crc kubenswrapper[4953]: I1203 00:13:57.358084 4953 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fe523418-5f1d-472a-bf22-246672d50480" path="/var/lib/kubelet/pods/fe523418-5f1d-472a-bf22-246672d50480/volumes" Dec 03 00:13:57 crc kubenswrapper[4953]: I1203 00:13:57.490023 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6c9bb59cb6-hxtzb" Dec 03 00:13:57 crc kubenswrapper[4953]: I1203 00:13:57.957082 4953 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-66df7c8f76-z74gr" Dec 03 00:13:57 crc kubenswrapper[4953]: I1203 00:13:57.985518 4953 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6c9bb59cb6-hxtzb"] Dec 03 00:13:57 crc kubenswrapper[4953]: W1203 00:13:57.992858 4953 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0b5291af_1bf9_43e2_a128_a421d69c67ea.slice/crio-d94518e04e74ee456eadec27caab338ed0c4719eaf2289f9a3e408d2d942828f WatchSource:0}: Error finding container d94518e04e74ee456eadec27caab338ed0c4719eaf2289f9a3e408d2d942828f: Status 404 returned error can't find the container with id d94518e04e74ee456eadec27caab338ed0c4719eaf2289f9a3e408d2d942828f Dec 03 00:13:58 crc kubenswrapper[4953]: I1203 00:13:58.030869 4953 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-t7ljb"] Dec 03 00:13:58 crc kubenswrapper[4953]: I1203 00:13:58.488535 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6c9bb59cb6-hxtzb" event={"ID":"0b5291af-1bf9-43e2-a128-a421d69c67ea","Type":"ContainerStarted","Data":"9fff4b9ea441dd96bc861c0936333b5e1a3192ba278f09fc40065561f75b82e2"} Dec 03 00:13:58 crc kubenswrapper[4953]: I1203 00:13:58.488600 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6c9bb59cb6-hxtzb" event={"ID":"0b5291af-1bf9-43e2-a128-a421d69c67ea","Type":"ContainerStarted","Data":"d94518e04e74ee456eadec27caab338ed0c4719eaf2289f9a3e408d2d942828f"} Dec 03 00:13:58 crc kubenswrapper[4953]: I1203 00:13:58.490343 4953 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6c9bb59cb6-hxtzb" Dec 03 00:13:58 crc kubenswrapper[4953]: I1203 00:13:58.589378 4953 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6c9bb59cb6-hxtzb" Dec 03 00:13:58 crc kubenswrapper[4953]: I1203 00:13:58.619615 4953 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6c9bb59cb6-hxtzb" podStartSLOduration=3.619593749 podStartE2EDuration="3.619593749s" podCreationTimestamp="2025-12-03 00:13:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 00:13:58.507922609 +0000 UTC m=+427.492203382" watchObservedRunningTime="2025-12-03 00:13:58.619593749 +0000 UTC m=+427.603874532" Dec 03 00:13:59 crc kubenswrapper[4953]: I1203 00:13:59.692666 4953 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-dpsq6" Dec 03 00:13:59 crc kubenswrapper[4953]: I1203 00:13:59.694288 4953 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-dpsq6" Dec 03 00:13:59 crc kubenswrapper[4953]: I1203 00:13:59.753726 4953 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-dpsq6" Dec 03 00:13:59 crc kubenswrapper[4953]: I1203 00:13:59.883067 4953 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-lcrlw" Dec 03 00:13:59 crc kubenswrapper[4953]: I1203 00:13:59.883163 4953 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-lcrlw" Dec 03 00:13:59 crc kubenswrapper[4953]: I1203 00:13:59.953221 4953 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-lcrlw" Dec 03 00:14:00 crc kubenswrapper[4953]: I1203 00:14:00.544987 4953 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-dpsq6" Dec 03 00:14:00 crc kubenswrapper[4953]: I1203 00:14:00.556932 4953 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-lcrlw" Dec 03 00:14:02 crc kubenswrapper[4953]: I1203 00:14:02.128596 4953 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-lxqn5" Dec 03 00:14:02 crc kubenswrapper[4953]: I1203 00:14:02.132684 4953 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-lxqn5" Dec 03 00:14:02 crc kubenswrapper[4953]: I1203 00:14:02.193703 4953 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-lxqn5" Dec 03 00:14:02 crc kubenswrapper[4953]: I1203 00:14:02.316689 4953 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-qrq2c" Dec 03 00:14:02 crc kubenswrapper[4953]: I1203 00:14:02.316751 4953 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-qrq2c" Dec 03 00:14:02 crc kubenswrapper[4953]: I1203 00:14:02.368674 4953 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-qrq2c" Dec 03 00:14:02 crc kubenswrapper[4953]: I1203 00:14:02.563088 4953 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-lxqn5" Dec 03 00:14:02 crc kubenswrapper[4953]: I1203 00:14:02.574521 4953 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-qrq2c" Dec 03 00:14:23 crc kubenswrapper[4953]: I1203 00:14:23.087114 4953 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-image-registry/image-registry-697d97f7c8-t7ljb" podUID="4313395c-8501-4b5b-a4cb-2fc986fced03" containerName="registry" containerID="cri-o://854d5d14e7ff3b9f049285ae27971f771bd934c5e0af219c59d73e018887a492" gracePeriod=30 Dec 03 00:14:23 crc kubenswrapper[4953]: I1203 00:14:23.528695 4953 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-t7ljb" Dec 03 00:14:23 crc kubenswrapper[4953]: I1203 00:14:23.601658 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/4313395c-8501-4b5b-a4cb-2fc986fced03-bound-sa-token\") pod \"4313395c-8501-4b5b-a4cb-2fc986fced03\" (UID: \"4313395c-8501-4b5b-a4cb-2fc986fced03\") " Dec 03 00:14:23 crc kubenswrapper[4953]: I1203 00:14:23.601739 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/4313395c-8501-4b5b-a4cb-2fc986fced03-registry-tls\") pod \"4313395c-8501-4b5b-a4cb-2fc986fced03\" (UID: \"4313395c-8501-4b5b-a4cb-2fc986fced03\") " Dec 03 00:14:23 crc kubenswrapper[4953]: I1203 00:14:23.601826 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/4313395c-8501-4b5b-a4cb-2fc986fced03-registry-certificates\") pod \"4313395c-8501-4b5b-a4cb-2fc986fced03\" (UID: \"4313395c-8501-4b5b-a4cb-2fc986fced03\") " Dec 03 00:14:23 crc kubenswrapper[4953]: I1203 00:14:23.601888 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c7jxx\" (UniqueName: \"kubernetes.io/projected/4313395c-8501-4b5b-a4cb-2fc986fced03-kube-api-access-c7jxx\") pod \"4313395c-8501-4b5b-a4cb-2fc986fced03\" (UID: \"4313395c-8501-4b5b-a4cb-2fc986fced03\") " Dec 03 00:14:23 crc kubenswrapper[4953]: I1203 00:14:23.601943 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/4313395c-8501-4b5b-a4cb-2fc986fced03-ca-trust-extracted\") pod \"4313395c-8501-4b5b-a4cb-2fc986fced03\" (UID: \"4313395c-8501-4b5b-a4cb-2fc986fced03\") " Dec 03 00:14:23 crc kubenswrapper[4953]: I1203 00:14:23.602028 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/4313395c-8501-4b5b-a4cb-2fc986fced03-trusted-ca\") pod \"4313395c-8501-4b5b-a4cb-2fc986fced03\" (UID: \"4313395c-8501-4b5b-a4cb-2fc986fced03\") " Dec 03 00:14:23 crc kubenswrapper[4953]: I1203 00:14:23.602057 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/4313395c-8501-4b5b-a4cb-2fc986fced03-installation-pull-secrets\") pod \"4313395c-8501-4b5b-a4cb-2fc986fced03\" (UID: \"4313395c-8501-4b5b-a4cb-2fc986fced03\") " Dec 03 00:14:23 crc kubenswrapper[4953]: I1203 00:14:23.603706 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4313395c-8501-4b5b-a4cb-2fc986fced03-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "4313395c-8501-4b5b-a4cb-2fc986fced03" (UID: "4313395c-8501-4b5b-a4cb-2fc986fced03"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:14:23 crc kubenswrapper[4953]: I1203 00:14:23.604317 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-storage\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"4313395c-8501-4b5b-a4cb-2fc986fced03\" (UID: \"4313395c-8501-4b5b-a4cb-2fc986fced03\") " Dec 03 00:14:23 crc kubenswrapper[4953]: I1203 00:14:23.603726 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4313395c-8501-4b5b-a4cb-2fc986fced03-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "4313395c-8501-4b5b-a4cb-2fc986fced03" (UID: "4313395c-8501-4b5b-a4cb-2fc986fced03"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:14:23 crc kubenswrapper[4953]: I1203 00:14:23.606105 4953 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/4313395c-8501-4b5b-a4cb-2fc986fced03-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 03 00:14:23 crc kubenswrapper[4953]: I1203 00:14:23.606145 4953 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/4313395c-8501-4b5b-a4cb-2fc986fced03-registry-certificates\") on node \"crc\" DevicePath \"\"" Dec 03 00:14:23 crc kubenswrapper[4953]: I1203 00:14:23.608080 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4313395c-8501-4b5b-a4cb-2fc986fced03-kube-api-access-c7jxx" (OuterVolumeSpecName: "kube-api-access-c7jxx") pod "4313395c-8501-4b5b-a4cb-2fc986fced03" (UID: "4313395c-8501-4b5b-a4cb-2fc986fced03"). InnerVolumeSpecName "kube-api-access-c7jxx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:14:23 crc kubenswrapper[4953]: I1203 00:14:23.608531 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4313395c-8501-4b5b-a4cb-2fc986fced03-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "4313395c-8501-4b5b-a4cb-2fc986fced03" (UID: "4313395c-8501-4b5b-a4cb-2fc986fced03"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:14:23 crc kubenswrapper[4953]: I1203 00:14:23.610008 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4313395c-8501-4b5b-a4cb-2fc986fced03-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "4313395c-8501-4b5b-a4cb-2fc986fced03" (UID: "4313395c-8501-4b5b-a4cb-2fc986fced03"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:14:23 crc kubenswrapper[4953]: I1203 00:14:23.611175 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4313395c-8501-4b5b-a4cb-2fc986fced03-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "4313395c-8501-4b5b-a4cb-2fc986fced03" (UID: "4313395c-8501-4b5b-a4cb-2fc986fced03"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:14:23 crc kubenswrapper[4953]: I1203 00:14:23.616791 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "registry-storage") pod "4313395c-8501-4b5b-a4cb-2fc986fced03" (UID: "4313395c-8501-4b5b-a4cb-2fc986fced03"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Dec 03 00:14:23 crc kubenswrapper[4953]: I1203 00:14:23.623314 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4313395c-8501-4b5b-a4cb-2fc986fced03-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "4313395c-8501-4b5b-a4cb-2fc986fced03" (UID: "4313395c-8501-4b5b-a4cb-2fc986fced03"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 00:14:23 crc kubenswrapper[4953]: I1203 00:14:23.642249 4953 generic.go:334] "Generic (PLEG): container finished" podID="4313395c-8501-4b5b-a4cb-2fc986fced03" containerID="854d5d14e7ff3b9f049285ae27971f771bd934c5e0af219c59d73e018887a492" exitCode=0 Dec 03 00:14:23 crc kubenswrapper[4953]: I1203 00:14:23.642316 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-t7ljb" event={"ID":"4313395c-8501-4b5b-a4cb-2fc986fced03","Type":"ContainerDied","Data":"854d5d14e7ff3b9f049285ae27971f771bd934c5e0af219c59d73e018887a492"} Dec 03 00:14:23 crc kubenswrapper[4953]: I1203 00:14:23.642361 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-t7ljb" event={"ID":"4313395c-8501-4b5b-a4cb-2fc986fced03","Type":"ContainerDied","Data":"61aa05ede5ab5d75673b721f1a77b40fa41eb1c55d50112cb6687d7cfa0d7305"} Dec 03 00:14:23 crc kubenswrapper[4953]: I1203 00:14:23.642385 4953 scope.go:117] "RemoveContainer" containerID="854d5d14e7ff3b9f049285ae27971f771bd934c5e0af219c59d73e018887a492" Dec 03 00:14:23 crc kubenswrapper[4953]: I1203 00:14:23.642318 4953 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-t7ljb" Dec 03 00:14:23 crc kubenswrapper[4953]: I1203 00:14:23.684342 4953 scope.go:117] "RemoveContainer" containerID="854d5d14e7ff3b9f049285ae27971f771bd934c5e0af219c59d73e018887a492" Dec 03 00:14:23 crc kubenswrapper[4953]: E1203 00:14:23.687138 4953 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"854d5d14e7ff3b9f049285ae27971f771bd934c5e0af219c59d73e018887a492\": container with ID starting with 854d5d14e7ff3b9f049285ae27971f771bd934c5e0af219c59d73e018887a492 not found: ID does not exist" containerID="854d5d14e7ff3b9f049285ae27971f771bd934c5e0af219c59d73e018887a492" Dec 03 00:14:23 crc kubenswrapper[4953]: I1203 00:14:23.687181 4953 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"854d5d14e7ff3b9f049285ae27971f771bd934c5e0af219c59d73e018887a492"} err="failed to get container status \"854d5d14e7ff3b9f049285ae27971f771bd934c5e0af219c59d73e018887a492\": rpc error: code = NotFound desc = could not find container \"854d5d14e7ff3b9f049285ae27971f771bd934c5e0af219c59d73e018887a492\": container with ID starting with 854d5d14e7ff3b9f049285ae27971f771bd934c5e0af219c59d73e018887a492 not found: ID does not exist" Dec 03 00:14:23 crc kubenswrapper[4953]: I1203 00:14:23.703675 4953 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-t7ljb"] Dec 03 00:14:23 crc kubenswrapper[4953]: I1203 00:14:23.707495 4953 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/4313395c-8501-4b5b-a4cb-2fc986fced03-bound-sa-token\") on node \"crc\" DevicePath \"\"" Dec 03 00:14:23 crc kubenswrapper[4953]: I1203 00:14:23.707528 4953 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/4313395c-8501-4b5b-a4cb-2fc986fced03-registry-tls\") on node \"crc\" DevicePath \"\"" Dec 03 00:14:23 crc kubenswrapper[4953]: I1203 00:14:23.707545 4953 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c7jxx\" (UniqueName: \"kubernetes.io/projected/4313395c-8501-4b5b-a4cb-2fc986fced03-kube-api-access-c7jxx\") on node \"crc\" DevicePath \"\"" Dec 03 00:14:23 crc kubenswrapper[4953]: I1203 00:14:23.707561 4953 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/4313395c-8501-4b5b-a4cb-2fc986fced03-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Dec 03 00:14:23 crc kubenswrapper[4953]: I1203 00:14:23.707573 4953 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/4313395c-8501-4b5b-a4cb-2fc986fced03-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Dec 03 00:14:23 crc kubenswrapper[4953]: I1203 00:14:23.709996 4953 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-t7ljb"] Dec 03 00:14:25 crc kubenswrapper[4953]: I1203 00:14:25.357122 4953 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4313395c-8501-4b5b-a4cb-2fc986fced03" path="/var/lib/kubelet/pods/4313395c-8501-4b5b-a4cb-2fc986fced03/volumes" Dec 03 00:15:00 crc kubenswrapper[4953]: I1203 00:15:00.200959 4953 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412015-b97kh"] Dec 03 00:15:00 crc kubenswrapper[4953]: E1203 00:15:00.201857 4953 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4313395c-8501-4b5b-a4cb-2fc986fced03" containerName="registry" Dec 03 00:15:00 crc kubenswrapper[4953]: I1203 00:15:00.201872 4953 state_mem.go:107] "Deleted CPUSet assignment" podUID="4313395c-8501-4b5b-a4cb-2fc986fced03" containerName="registry" Dec 03 00:15:00 crc kubenswrapper[4953]: I1203 00:15:00.202016 4953 memory_manager.go:354] "RemoveStaleState removing state" podUID="4313395c-8501-4b5b-a4cb-2fc986fced03" containerName="registry" Dec 03 00:15:00 crc kubenswrapper[4953]: I1203 00:15:00.202479 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29412015-b97kh" Dec 03 00:15:00 crc kubenswrapper[4953]: I1203 00:15:00.204810 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 03 00:15:00 crc kubenswrapper[4953]: I1203 00:15:00.205835 4953 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 03 00:15:00 crc kubenswrapper[4953]: I1203 00:15:00.215471 4953 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412015-b97kh"] Dec 03 00:15:00 crc kubenswrapper[4953]: I1203 00:15:00.254578 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/879c4173-81ae-4470-a852-57cd81af3d24-secret-volume\") pod \"collect-profiles-29412015-b97kh\" (UID: \"879c4173-81ae-4470-a852-57cd81af3d24\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412015-b97kh" Dec 03 00:15:00 crc kubenswrapper[4953]: I1203 00:15:00.254716 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/879c4173-81ae-4470-a852-57cd81af3d24-config-volume\") pod \"collect-profiles-29412015-b97kh\" (UID: \"879c4173-81ae-4470-a852-57cd81af3d24\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412015-b97kh" Dec 03 00:15:00 crc kubenswrapper[4953]: I1203 00:15:00.254758 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fdt94\" (UniqueName: \"kubernetes.io/projected/879c4173-81ae-4470-a852-57cd81af3d24-kube-api-access-fdt94\") pod \"collect-profiles-29412015-b97kh\" (UID: \"879c4173-81ae-4470-a852-57cd81af3d24\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412015-b97kh" Dec 03 00:15:00 crc kubenswrapper[4953]: I1203 00:15:00.356817 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/879c4173-81ae-4470-a852-57cd81af3d24-secret-volume\") pod \"collect-profiles-29412015-b97kh\" (UID: \"879c4173-81ae-4470-a852-57cd81af3d24\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412015-b97kh" Dec 03 00:15:00 crc kubenswrapper[4953]: I1203 00:15:00.356930 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/879c4173-81ae-4470-a852-57cd81af3d24-config-volume\") pod \"collect-profiles-29412015-b97kh\" (UID: \"879c4173-81ae-4470-a852-57cd81af3d24\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412015-b97kh" Dec 03 00:15:00 crc kubenswrapper[4953]: I1203 00:15:00.357025 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fdt94\" (UniqueName: \"kubernetes.io/projected/879c4173-81ae-4470-a852-57cd81af3d24-kube-api-access-fdt94\") pod \"collect-profiles-29412015-b97kh\" (UID: \"879c4173-81ae-4470-a852-57cd81af3d24\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412015-b97kh" Dec 03 00:15:00 crc kubenswrapper[4953]: I1203 00:15:00.358195 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/879c4173-81ae-4470-a852-57cd81af3d24-config-volume\") pod \"collect-profiles-29412015-b97kh\" (UID: \"879c4173-81ae-4470-a852-57cd81af3d24\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412015-b97kh" Dec 03 00:15:00 crc kubenswrapper[4953]: I1203 00:15:00.369700 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/879c4173-81ae-4470-a852-57cd81af3d24-secret-volume\") pod \"collect-profiles-29412015-b97kh\" (UID: \"879c4173-81ae-4470-a852-57cd81af3d24\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412015-b97kh" Dec 03 00:15:00 crc kubenswrapper[4953]: I1203 00:15:00.376848 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fdt94\" (UniqueName: \"kubernetes.io/projected/879c4173-81ae-4470-a852-57cd81af3d24-kube-api-access-fdt94\") pod \"collect-profiles-29412015-b97kh\" (UID: \"879c4173-81ae-4470-a852-57cd81af3d24\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412015-b97kh" Dec 03 00:15:00 crc kubenswrapper[4953]: I1203 00:15:00.524255 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29412015-b97kh" Dec 03 00:15:00 crc kubenswrapper[4953]: I1203 00:15:00.753365 4953 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412015-b97kh"] Dec 03 00:15:00 crc kubenswrapper[4953]: I1203 00:15:00.871666 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29412015-b97kh" event={"ID":"879c4173-81ae-4470-a852-57cd81af3d24","Type":"ContainerStarted","Data":"22d0f31e5056dc6affd356347876811a9cb442c4ed320aa943381175be525951"} Dec 03 00:15:01 crc kubenswrapper[4953]: I1203 00:15:01.880332 4953 generic.go:334] "Generic (PLEG): container finished" podID="879c4173-81ae-4470-a852-57cd81af3d24" containerID="d64114c2085964dbd099d75399c5cb42135c0c36e6a3919ef16aef78bfbc13ce" exitCode=0 Dec 03 00:15:01 crc kubenswrapper[4953]: I1203 00:15:01.880437 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29412015-b97kh" event={"ID":"879c4173-81ae-4470-a852-57cd81af3d24","Type":"ContainerDied","Data":"d64114c2085964dbd099d75399c5cb42135c0c36e6a3919ef16aef78bfbc13ce"} Dec 03 00:15:03 crc kubenswrapper[4953]: I1203 00:15:03.177277 4953 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29412015-b97kh" Dec 03 00:15:03 crc kubenswrapper[4953]: I1203 00:15:03.197153 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fdt94\" (UniqueName: \"kubernetes.io/projected/879c4173-81ae-4470-a852-57cd81af3d24-kube-api-access-fdt94\") pod \"879c4173-81ae-4470-a852-57cd81af3d24\" (UID: \"879c4173-81ae-4470-a852-57cd81af3d24\") " Dec 03 00:15:03 crc kubenswrapper[4953]: I1203 00:15:03.197256 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/879c4173-81ae-4470-a852-57cd81af3d24-config-volume\") pod \"879c4173-81ae-4470-a852-57cd81af3d24\" (UID: \"879c4173-81ae-4470-a852-57cd81af3d24\") " Dec 03 00:15:03 crc kubenswrapper[4953]: I1203 00:15:03.198885 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/879c4173-81ae-4470-a852-57cd81af3d24-config-volume" (OuterVolumeSpecName: "config-volume") pod "879c4173-81ae-4470-a852-57cd81af3d24" (UID: "879c4173-81ae-4470-a852-57cd81af3d24"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:15:03 crc kubenswrapper[4953]: I1203 00:15:03.204384 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/879c4173-81ae-4470-a852-57cd81af3d24-kube-api-access-fdt94" (OuterVolumeSpecName: "kube-api-access-fdt94") pod "879c4173-81ae-4470-a852-57cd81af3d24" (UID: "879c4173-81ae-4470-a852-57cd81af3d24"). InnerVolumeSpecName "kube-api-access-fdt94". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:15:03 crc kubenswrapper[4953]: I1203 00:15:03.298443 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/879c4173-81ae-4470-a852-57cd81af3d24-secret-volume\") pod \"879c4173-81ae-4470-a852-57cd81af3d24\" (UID: \"879c4173-81ae-4470-a852-57cd81af3d24\") " Dec 03 00:15:03 crc kubenswrapper[4953]: I1203 00:15:03.298789 4953 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fdt94\" (UniqueName: \"kubernetes.io/projected/879c4173-81ae-4470-a852-57cd81af3d24-kube-api-access-fdt94\") on node \"crc\" DevicePath \"\"" Dec 03 00:15:03 crc kubenswrapper[4953]: I1203 00:15:03.298805 4953 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/879c4173-81ae-4470-a852-57cd81af3d24-config-volume\") on node \"crc\" DevicePath \"\"" Dec 03 00:15:03 crc kubenswrapper[4953]: I1203 00:15:03.302807 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/879c4173-81ae-4470-a852-57cd81af3d24-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "879c4173-81ae-4470-a852-57cd81af3d24" (UID: "879c4173-81ae-4470-a852-57cd81af3d24"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:15:03 crc kubenswrapper[4953]: I1203 00:15:03.399840 4953 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/879c4173-81ae-4470-a852-57cd81af3d24-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 03 00:15:03 crc kubenswrapper[4953]: I1203 00:15:03.895565 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29412015-b97kh" event={"ID":"879c4173-81ae-4470-a852-57cd81af3d24","Type":"ContainerDied","Data":"22d0f31e5056dc6affd356347876811a9cb442c4ed320aa943381175be525951"} Dec 03 00:15:03 crc kubenswrapper[4953]: I1203 00:15:03.895607 4953 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="22d0f31e5056dc6affd356347876811a9cb442c4ed320aa943381175be525951" Dec 03 00:15:03 crc kubenswrapper[4953]: I1203 00:15:03.895711 4953 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29412015-b97kh" Dec 03 00:16:18 crc kubenswrapper[4953]: I1203 00:16:18.945213 4953 patch_prober.go:28] interesting pod/machine-config-daemon-6dfbn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 00:16:18 crc kubenswrapper[4953]: I1203 00:16:18.946201 4953 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-6dfbn" podUID="53ffb1e9-9a2e-4869-80d7-776f4aa7ef8b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 00:16:48 crc kubenswrapper[4953]: I1203 00:16:48.944712 4953 patch_prober.go:28] interesting pod/machine-config-daemon-6dfbn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 00:16:48 crc kubenswrapper[4953]: I1203 00:16:48.945810 4953 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-6dfbn" podUID="53ffb1e9-9a2e-4869-80d7-776f4aa7ef8b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 00:17:18 crc kubenswrapper[4953]: I1203 00:17:18.944504 4953 patch_prober.go:28] interesting pod/machine-config-daemon-6dfbn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 00:17:18 crc kubenswrapper[4953]: I1203 00:17:18.945788 4953 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-6dfbn" podUID="53ffb1e9-9a2e-4869-80d7-776f4aa7ef8b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 00:17:18 crc kubenswrapper[4953]: I1203 00:17:18.945865 4953 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-6dfbn" Dec 03 00:17:18 crc kubenswrapper[4953]: I1203 00:17:18.946898 4953 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"257c3b050511c36fa41f2b2785b39713a95149ac43905452ee5690643ac49dc2"} pod="openshift-machine-config-operator/machine-config-daemon-6dfbn" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 03 00:17:18 crc kubenswrapper[4953]: I1203 00:17:18.947003 4953 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-6dfbn" podUID="53ffb1e9-9a2e-4869-80d7-776f4aa7ef8b" containerName="machine-config-daemon" containerID="cri-o://257c3b050511c36fa41f2b2785b39713a95149ac43905452ee5690643ac49dc2" gracePeriod=600 Dec 03 00:17:19 crc kubenswrapper[4953]: I1203 00:17:19.817702 4953 generic.go:334] "Generic (PLEG): container finished" podID="53ffb1e9-9a2e-4869-80d7-776f4aa7ef8b" containerID="257c3b050511c36fa41f2b2785b39713a95149ac43905452ee5690643ac49dc2" exitCode=0 Dec 03 00:17:19 crc kubenswrapper[4953]: I1203 00:17:19.817769 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-6dfbn" event={"ID":"53ffb1e9-9a2e-4869-80d7-776f4aa7ef8b","Type":"ContainerDied","Data":"257c3b050511c36fa41f2b2785b39713a95149ac43905452ee5690643ac49dc2"} Dec 03 00:17:19 crc kubenswrapper[4953]: I1203 00:17:19.818764 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-6dfbn" event={"ID":"53ffb1e9-9a2e-4869-80d7-776f4aa7ef8b","Type":"ContainerStarted","Data":"553caefdc3000cd9d8eaf08bf4d23cdb2a37135c2cd1053fce9b7f6cb2b25c0b"} Dec 03 00:17:19 crc kubenswrapper[4953]: I1203 00:17:19.818793 4953 scope.go:117] "RemoveContainer" containerID="cd3fdc5de406f82703cc48138f69373c50d1862f3fb1f4cc4572aa863d2eca4c" Dec 03 00:19:10 crc kubenswrapper[4953]: I1203 00:19:10.464759 4953 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-txffx"] Dec 03 00:19:10 crc kubenswrapper[4953]: I1203 00:19:10.472055 4953 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-txffx" podUID="8816ebee-c735-4b8d-b1ec-65349707b347" containerName="ovn-controller" containerID="cri-o://dc0a5d4cc9d614d8492626a1b114bdc52f510eb5965d842c02c083f08d4849a7" gracePeriod=30 Dec 03 00:19:10 crc kubenswrapper[4953]: I1203 00:19:10.472137 4953 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-txffx" podUID="8816ebee-c735-4b8d-b1ec-65349707b347" containerName="sbdb" containerID="cri-o://7deadc8e135ff1bc7c5a2f093e489bb5b4b6f166bc874c6acc2a841c5397f11e" gracePeriod=30 Dec 03 00:19:10 crc kubenswrapper[4953]: I1203 00:19:10.472242 4953 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-txffx" podUID="8816ebee-c735-4b8d-b1ec-65349707b347" containerName="northd" containerID="cri-o://fd5f428de229749485267e81d34e31b708728b52474032e88460e54747f21e11" gracePeriod=30 Dec 03 00:19:10 crc kubenswrapper[4953]: I1203 00:19:10.472313 4953 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-txffx" podUID="8816ebee-c735-4b8d-b1ec-65349707b347" containerName="kube-rbac-proxy-ovn-metrics" containerID="cri-o://e0f6f645a58993f052a6f365bb808f62d99f7fea74886c5cc022b2d86112321a" gracePeriod=30 Dec 03 00:19:10 crc kubenswrapper[4953]: I1203 00:19:10.472347 4953 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-txffx" podUID="8816ebee-c735-4b8d-b1ec-65349707b347" containerName="kube-rbac-proxy-node" containerID="cri-o://d4b85224419a4075acb340e4b11095f2adb5c316e285afa4e19295c2181c0d24" gracePeriod=30 Dec 03 00:19:10 crc kubenswrapper[4953]: I1203 00:19:10.472223 4953 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-txffx" podUID="8816ebee-c735-4b8d-b1ec-65349707b347" containerName="nbdb" containerID="cri-o://a61b951a9cbab2c48c0f2df8351f97d7ff71bf3674f641fe5eac30155232b224" gracePeriod=30 Dec 03 00:19:10 crc kubenswrapper[4953]: I1203 00:19:10.472406 4953 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-txffx" podUID="8816ebee-c735-4b8d-b1ec-65349707b347" containerName="ovn-acl-logging" containerID="cri-o://00a2ea4a9bb870fce30e7fa845f787f184e7a19e42c95511655ea8c5f84a26d1" gracePeriod=30 Dec 03 00:19:10 crc kubenswrapper[4953]: I1203 00:19:10.526574 4953 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-txffx" podUID="8816ebee-c735-4b8d-b1ec-65349707b347" containerName="ovnkube-controller" containerID="cri-o://bcde8ca2a20e2afee0f6cb7164e6068aa82f05c0c978ea3cc4c23618fcb57bac" gracePeriod=30 Dec 03 00:19:10 crc kubenswrapper[4953]: I1203 00:19:10.559932 4953 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-675qk_43a4bad8-71f4-441a-8302-aeb526a986af/kube-multus/2.log" Dec 03 00:19:10 crc kubenswrapper[4953]: I1203 00:19:10.560790 4953 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-675qk_43a4bad8-71f4-441a-8302-aeb526a986af/kube-multus/1.log" Dec 03 00:19:10 crc kubenswrapper[4953]: I1203 00:19:10.560827 4953 generic.go:334] "Generic (PLEG): container finished" podID="43a4bad8-71f4-441a-8302-aeb526a986af" containerID="599a93b23125c53485f61428a83e40ff9255f5e0e3ce3f5f2c0400cbeede0384" exitCode=2 Dec 03 00:19:10 crc kubenswrapper[4953]: I1203 00:19:10.560856 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-675qk" event={"ID":"43a4bad8-71f4-441a-8302-aeb526a986af","Type":"ContainerDied","Data":"599a93b23125c53485f61428a83e40ff9255f5e0e3ce3f5f2c0400cbeede0384"} Dec 03 00:19:10 crc kubenswrapper[4953]: I1203 00:19:10.560889 4953 scope.go:117] "RemoveContainer" containerID="e701c2ad05014683415ae2fb03b96f3acb38062ba7c5bfc4d0a9e34431695e36" Dec 03 00:19:10 crc kubenswrapper[4953]: I1203 00:19:10.562531 4953 scope.go:117] "RemoveContainer" containerID="599a93b23125c53485f61428a83e40ff9255f5e0e3ce3f5f2c0400cbeede0384" Dec 03 00:19:10 crc kubenswrapper[4953]: E1203 00:19:10.599235 4953 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8816ebee_c735_4b8d_b1ec_65349707b347.slice/crio-conmon-d4b85224419a4075acb340e4b11095f2adb5c316e285afa4e19295c2181c0d24.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8816ebee_c735_4b8d_b1ec_65349707b347.slice/crio-conmon-dc0a5d4cc9d614d8492626a1b114bdc52f510eb5965d842c02c083f08d4849a7.scope\": RecentStats: unable to find data in memory cache]" Dec 03 00:19:10 crc kubenswrapper[4953]: I1203 00:19:10.846480 4953 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-txffx_8816ebee-c735-4b8d-b1ec-65349707b347/ovnkube-controller/3.log" Dec 03 00:19:10 crc kubenswrapper[4953]: I1203 00:19:10.850071 4953 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-txffx_8816ebee-c735-4b8d-b1ec-65349707b347/ovn-acl-logging/0.log" Dec 03 00:19:10 crc kubenswrapper[4953]: I1203 00:19:10.850656 4953 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-txffx_8816ebee-c735-4b8d-b1ec-65349707b347/ovn-controller/0.log" Dec 03 00:19:10 crc kubenswrapper[4953]: I1203 00:19:10.851246 4953 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-txffx" Dec 03 00:19:10 crc kubenswrapper[4953]: I1203 00:19:10.916507 4953 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-fnl4b"] Dec 03 00:19:10 crc kubenswrapper[4953]: E1203 00:19:10.916780 4953 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8816ebee-c735-4b8d-b1ec-65349707b347" containerName="nbdb" Dec 03 00:19:10 crc kubenswrapper[4953]: I1203 00:19:10.916796 4953 state_mem.go:107] "Deleted CPUSet assignment" podUID="8816ebee-c735-4b8d-b1ec-65349707b347" containerName="nbdb" Dec 03 00:19:10 crc kubenswrapper[4953]: E1203 00:19:10.916811 4953 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8816ebee-c735-4b8d-b1ec-65349707b347" containerName="ovn-acl-logging" Dec 03 00:19:10 crc kubenswrapper[4953]: I1203 00:19:10.916818 4953 state_mem.go:107] "Deleted CPUSet assignment" podUID="8816ebee-c735-4b8d-b1ec-65349707b347" containerName="ovn-acl-logging" Dec 03 00:19:10 crc kubenswrapper[4953]: E1203 00:19:10.916830 4953 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8816ebee-c735-4b8d-b1ec-65349707b347" containerName="sbdb" Dec 03 00:19:10 crc kubenswrapper[4953]: I1203 00:19:10.916836 4953 state_mem.go:107] "Deleted CPUSet assignment" podUID="8816ebee-c735-4b8d-b1ec-65349707b347" containerName="sbdb" Dec 03 00:19:10 crc kubenswrapper[4953]: E1203 00:19:10.916847 4953 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8816ebee-c735-4b8d-b1ec-65349707b347" containerName="northd" Dec 03 00:19:10 crc kubenswrapper[4953]: I1203 00:19:10.916857 4953 state_mem.go:107] "Deleted CPUSet assignment" podUID="8816ebee-c735-4b8d-b1ec-65349707b347" containerName="northd" Dec 03 00:19:10 crc kubenswrapper[4953]: E1203 00:19:10.916865 4953 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="879c4173-81ae-4470-a852-57cd81af3d24" containerName="collect-profiles" Dec 03 00:19:10 crc kubenswrapper[4953]: I1203 00:19:10.916871 4953 state_mem.go:107] "Deleted CPUSet assignment" podUID="879c4173-81ae-4470-a852-57cd81af3d24" containerName="collect-profiles" Dec 03 00:19:10 crc kubenswrapper[4953]: E1203 00:19:10.916877 4953 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8816ebee-c735-4b8d-b1ec-65349707b347" containerName="kube-rbac-proxy-ovn-metrics" Dec 03 00:19:10 crc kubenswrapper[4953]: I1203 00:19:10.916884 4953 state_mem.go:107] "Deleted CPUSet assignment" podUID="8816ebee-c735-4b8d-b1ec-65349707b347" containerName="kube-rbac-proxy-ovn-metrics" Dec 03 00:19:10 crc kubenswrapper[4953]: E1203 00:19:10.916892 4953 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8816ebee-c735-4b8d-b1ec-65349707b347" containerName="ovnkube-controller" Dec 03 00:19:10 crc kubenswrapper[4953]: I1203 00:19:10.916900 4953 state_mem.go:107] "Deleted CPUSet assignment" podUID="8816ebee-c735-4b8d-b1ec-65349707b347" containerName="ovnkube-controller" Dec 03 00:19:10 crc kubenswrapper[4953]: E1203 00:19:10.916908 4953 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8816ebee-c735-4b8d-b1ec-65349707b347" containerName="ovnkube-controller" Dec 03 00:19:10 crc kubenswrapper[4953]: I1203 00:19:10.916914 4953 state_mem.go:107] "Deleted CPUSet assignment" podUID="8816ebee-c735-4b8d-b1ec-65349707b347" containerName="ovnkube-controller" Dec 03 00:19:10 crc kubenswrapper[4953]: E1203 00:19:10.916923 4953 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8816ebee-c735-4b8d-b1ec-65349707b347" containerName="ovnkube-controller" Dec 03 00:19:10 crc kubenswrapper[4953]: I1203 00:19:10.916928 4953 state_mem.go:107] "Deleted CPUSet assignment" podUID="8816ebee-c735-4b8d-b1ec-65349707b347" containerName="ovnkube-controller" Dec 03 00:19:10 crc kubenswrapper[4953]: E1203 00:19:10.916936 4953 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8816ebee-c735-4b8d-b1ec-65349707b347" containerName="kubecfg-setup" Dec 03 00:19:10 crc kubenswrapper[4953]: I1203 00:19:10.916942 4953 state_mem.go:107] "Deleted CPUSet assignment" podUID="8816ebee-c735-4b8d-b1ec-65349707b347" containerName="kubecfg-setup" Dec 03 00:19:10 crc kubenswrapper[4953]: E1203 00:19:10.916952 4953 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8816ebee-c735-4b8d-b1ec-65349707b347" containerName="kube-rbac-proxy-node" Dec 03 00:19:10 crc kubenswrapper[4953]: I1203 00:19:10.916959 4953 state_mem.go:107] "Deleted CPUSet assignment" podUID="8816ebee-c735-4b8d-b1ec-65349707b347" containerName="kube-rbac-proxy-node" Dec 03 00:19:10 crc kubenswrapper[4953]: E1203 00:19:10.917008 4953 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8816ebee-c735-4b8d-b1ec-65349707b347" containerName="ovnkube-controller" Dec 03 00:19:10 crc kubenswrapper[4953]: I1203 00:19:10.917018 4953 state_mem.go:107] "Deleted CPUSet assignment" podUID="8816ebee-c735-4b8d-b1ec-65349707b347" containerName="ovnkube-controller" Dec 03 00:19:10 crc kubenswrapper[4953]: E1203 00:19:10.917032 4953 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8816ebee-c735-4b8d-b1ec-65349707b347" containerName="ovn-controller" Dec 03 00:19:10 crc kubenswrapper[4953]: I1203 00:19:10.917042 4953 state_mem.go:107] "Deleted CPUSet assignment" podUID="8816ebee-c735-4b8d-b1ec-65349707b347" containerName="ovn-controller" Dec 03 00:19:10 crc kubenswrapper[4953]: I1203 00:19:10.917155 4953 memory_manager.go:354] "RemoveStaleState removing state" podUID="8816ebee-c735-4b8d-b1ec-65349707b347" containerName="ovnkube-controller" Dec 03 00:19:10 crc kubenswrapper[4953]: I1203 00:19:10.917168 4953 memory_manager.go:354] "RemoveStaleState removing state" podUID="8816ebee-c735-4b8d-b1ec-65349707b347" containerName="ovnkube-controller" Dec 03 00:19:10 crc kubenswrapper[4953]: I1203 00:19:10.917181 4953 memory_manager.go:354] "RemoveStaleState removing state" podUID="8816ebee-c735-4b8d-b1ec-65349707b347" containerName="ovnkube-controller" Dec 03 00:19:10 crc kubenswrapper[4953]: I1203 00:19:10.917189 4953 memory_manager.go:354] "RemoveStaleState removing state" podUID="8816ebee-c735-4b8d-b1ec-65349707b347" containerName="northd" Dec 03 00:19:10 crc kubenswrapper[4953]: I1203 00:19:10.917198 4953 memory_manager.go:354] "RemoveStaleState removing state" podUID="8816ebee-c735-4b8d-b1ec-65349707b347" containerName="ovnkube-controller" Dec 03 00:19:10 crc kubenswrapper[4953]: I1203 00:19:10.917208 4953 memory_manager.go:354] "RemoveStaleState removing state" podUID="8816ebee-c735-4b8d-b1ec-65349707b347" containerName="sbdb" Dec 03 00:19:10 crc kubenswrapper[4953]: I1203 00:19:10.917217 4953 memory_manager.go:354] "RemoveStaleState removing state" podUID="8816ebee-c735-4b8d-b1ec-65349707b347" containerName="ovn-controller" Dec 03 00:19:10 crc kubenswrapper[4953]: I1203 00:19:10.917226 4953 memory_manager.go:354] "RemoveStaleState removing state" podUID="8816ebee-c735-4b8d-b1ec-65349707b347" containerName="kube-rbac-proxy-node" Dec 03 00:19:10 crc kubenswrapper[4953]: I1203 00:19:10.917234 4953 memory_manager.go:354] "RemoveStaleState removing state" podUID="879c4173-81ae-4470-a852-57cd81af3d24" containerName="collect-profiles" Dec 03 00:19:10 crc kubenswrapper[4953]: I1203 00:19:10.917242 4953 memory_manager.go:354] "RemoveStaleState removing state" podUID="8816ebee-c735-4b8d-b1ec-65349707b347" containerName="ovn-acl-logging" Dec 03 00:19:10 crc kubenswrapper[4953]: I1203 00:19:10.917249 4953 memory_manager.go:354] "RemoveStaleState removing state" podUID="8816ebee-c735-4b8d-b1ec-65349707b347" containerName="kube-rbac-proxy-ovn-metrics" Dec 03 00:19:10 crc kubenswrapper[4953]: I1203 00:19:10.917257 4953 memory_manager.go:354] "RemoveStaleState removing state" podUID="8816ebee-c735-4b8d-b1ec-65349707b347" containerName="nbdb" Dec 03 00:19:10 crc kubenswrapper[4953]: E1203 00:19:10.917372 4953 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8816ebee-c735-4b8d-b1ec-65349707b347" containerName="ovnkube-controller" Dec 03 00:19:10 crc kubenswrapper[4953]: I1203 00:19:10.917379 4953 state_mem.go:107] "Deleted CPUSet assignment" podUID="8816ebee-c735-4b8d-b1ec-65349707b347" containerName="ovnkube-controller" Dec 03 00:19:10 crc kubenswrapper[4953]: I1203 00:19:10.917477 4953 memory_manager.go:354] "RemoveStaleState removing state" podUID="8816ebee-c735-4b8d-b1ec-65349707b347" containerName="ovnkube-controller" Dec 03 00:19:10 crc kubenswrapper[4953]: I1203 00:19:10.919251 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-fnl4b" Dec 03 00:19:11 crc kubenswrapper[4953]: I1203 00:19:11.030544 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/8816ebee-c735-4b8d-b1ec-65349707b347-host-run-ovn-kubernetes\") pod \"8816ebee-c735-4b8d-b1ec-65349707b347\" (UID: \"8816ebee-c735-4b8d-b1ec-65349707b347\") " Dec 03 00:19:11 crc kubenswrapper[4953]: I1203 00:19:11.030616 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/8816ebee-c735-4b8d-b1ec-65349707b347-var-lib-openvswitch\") pod \"8816ebee-c735-4b8d-b1ec-65349707b347\" (UID: \"8816ebee-c735-4b8d-b1ec-65349707b347\") " Dec 03 00:19:11 crc kubenswrapper[4953]: I1203 00:19:11.030660 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/8816ebee-c735-4b8d-b1ec-65349707b347-run-ovn\") pod \"8816ebee-c735-4b8d-b1ec-65349707b347\" (UID: \"8816ebee-c735-4b8d-b1ec-65349707b347\") " Dec 03 00:19:11 crc kubenswrapper[4953]: I1203 00:19:11.030696 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m2fdf\" (UniqueName: \"kubernetes.io/projected/8816ebee-c735-4b8d-b1ec-65349707b347-kube-api-access-m2fdf\") pod \"8816ebee-c735-4b8d-b1ec-65349707b347\" (UID: \"8816ebee-c735-4b8d-b1ec-65349707b347\") " Dec 03 00:19:11 crc kubenswrapper[4953]: I1203 00:19:11.030721 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/8816ebee-c735-4b8d-b1ec-65349707b347-host-kubelet\") pod \"8816ebee-c735-4b8d-b1ec-65349707b347\" (UID: \"8816ebee-c735-4b8d-b1ec-65349707b347\") " Dec 03 00:19:11 crc kubenswrapper[4953]: I1203 00:19:11.030743 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/8816ebee-c735-4b8d-b1ec-65349707b347-host-run-netns\") pod \"8816ebee-c735-4b8d-b1ec-65349707b347\" (UID: \"8816ebee-c735-4b8d-b1ec-65349707b347\") " Dec 03 00:19:11 crc kubenswrapper[4953]: I1203 00:19:11.030724 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/8816ebee-c735-4b8d-b1ec-65349707b347-host-run-ovn-kubernetes" (OuterVolumeSpecName: "host-run-ovn-kubernetes") pod "8816ebee-c735-4b8d-b1ec-65349707b347" (UID: "8816ebee-c735-4b8d-b1ec-65349707b347"). InnerVolumeSpecName "host-run-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 00:19:11 crc kubenswrapper[4953]: I1203 00:19:11.030777 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/8816ebee-c735-4b8d-b1ec-65349707b347-node-log\") pod \"8816ebee-c735-4b8d-b1ec-65349707b347\" (UID: \"8816ebee-c735-4b8d-b1ec-65349707b347\") " Dec 03 00:19:11 crc kubenswrapper[4953]: I1203 00:19:11.030769 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/8816ebee-c735-4b8d-b1ec-65349707b347-run-ovn" (OuterVolumeSpecName: "run-ovn") pod "8816ebee-c735-4b8d-b1ec-65349707b347" (UID: "8816ebee-c735-4b8d-b1ec-65349707b347"). InnerVolumeSpecName "run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 00:19:11 crc kubenswrapper[4953]: I1203 00:19:11.030811 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/8816ebee-c735-4b8d-b1ec-65349707b347-env-overrides\") pod \"8816ebee-c735-4b8d-b1ec-65349707b347\" (UID: \"8816ebee-c735-4b8d-b1ec-65349707b347\") " Dec 03 00:19:11 crc kubenswrapper[4953]: I1203 00:19:11.030814 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/8816ebee-c735-4b8d-b1ec-65349707b347-host-kubelet" (OuterVolumeSpecName: "host-kubelet") pod "8816ebee-c735-4b8d-b1ec-65349707b347" (UID: "8816ebee-c735-4b8d-b1ec-65349707b347"). InnerVolumeSpecName "host-kubelet". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 00:19:11 crc kubenswrapper[4953]: I1203 00:19:11.030837 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/8816ebee-c735-4b8d-b1ec-65349707b347-host-cni-bin\") pod \"8816ebee-c735-4b8d-b1ec-65349707b347\" (UID: \"8816ebee-c735-4b8d-b1ec-65349707b347\") " Dec 03 00:19:11 crc kubenswrapper[4953]: I1203 00:19:11.030839 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/8816ebee-c735-4b8d-b1ec-65349707b347-host-run-netns" (OuterVolumeSpecName: "host-run-netns") pod "8816ebee-c735-4b8d-b1ec-65349707b347" (UID: "8816ebee-c735-4b8d-b1ec-65349707b347"). InnerVolumeSpecName "host-run-netns". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 00:19:11 crc kubenswrapper[4953]: I1203 00:19:11.030863 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/8816ebee-c735-4b8d-b1ec-65349707b347-node-log" (OuterVolumeSpecName: "node-log") pod "8816ebee-c735-4b8d-b1ec-65349707b347" (UID: "8816ebee-c735-4b8d-b1ec-65349707b347"). InnerVolumeSpecName "node-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 00:19:11 crc kubenswrapper[4953]: I1203 00:19:11.030769 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/8816ebee-c735-4b8d-b1ec-65349707b347-var-lib-openvswitch" (OuterVolumeSpecName: "var-lib-openvswitch") pod "8816ebee-c735-4b8d-b1ec-65349707b347" (UID: "8816ebee-c735-4b8d-b1ec-65349707b347"). InnerVolumeSpecName "var-lib-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 00:19:11 crc kubenswrapper[4953]: I1203 00:19:11.030871 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/8816ebee-c735-4b8d-b1ec-65349707b347-log-socket\") pod \"8816ebee-c735-4b8d-b1ec-65349707b347\" (UID: \"8816ebee-c735-4b8d-b1ec-65349707b347\") " Dec 03 00:19:11 crc kubenswrapper[4953]: I1203 00:19:11.030892 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/8816ebee-c735-4b8d-b1ec-65349707b347-host-cni-bin" (OuterVolumeSpecName: "host-cni-bin") pod "8816ebee-c735-4b8d-b1ec-65349707b347" (UID: "8816ebee-c735-4b8d-b1ec-65349707b347"). InnerVolumeSpecName "host-cni-bin". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 00:19:11 crc kubenswrapper[4953]: I1203 00:19:11.030949 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/8816ebee-c735-4b8d-b1ec-65349707b347-log-socket" (OuterVolumeSpecName: "log-socket") pod "8816ebee-c735-4b8d-b1ec-65349707b347" (UID: "8816ebee-c735-4b8d-b1ec-65349707b347"). InnerVolumeSpecName "log-socket". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 00:19:11 crc kubenswrapper[4953]: I1203 00:19:11.030950 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/8816ebee-c735-4b8d-b1ec-65349707b347-etc-openvswitch\") pod \"8816ebee-c735-4b8d-b1ec-65349707b347\" (UID: \"8816ebee-c735-4b8d-b1ec-65349707b347\") " Dec 03 00:19:11 crc kubenswrapper[4953]: I1203 00:19:11.031012 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/8816ebee-c735-4b8d-b1ec-65349707b347-etc-openvswitch" (OuterVolumeSpecName: "etc-openvswitch") pod "8816ebee-c735-4b8d-b1ec-65349707b347" (UID: "8816ebee-c735-4b8d-b1ec-65349707b347"). InnerVolumeSpecName "etc-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 00:19:11 crc kubenswrapper[4953]: I1203 00:19:11.031049 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/8816ebee-c735-4b8d-b1ec-65349707b347-host-var-lib-cni-networks-ovn-kubernetes\") pod \"8816ebee-c735-4b8d-b1ec-65349707b347\" (UID: \"8816ebee-c735-4b8d-b1ec-65349707b347\") " Dec 03 00:19:11 crc kubenswrapper[4953]: I1203 00:19:11.031080 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/8816ebee-c735-4b8d-b1ec-65349707b347-host-cni-netd\") pod \"8816ebee-c735-4b8d-b1ec-65349707b347\" (UID: \"8816ebee-c735-4b8d-b1ec-65349707b347\") " Dec 03 00:19:11 crc kubenswrapper[4953]: I1203 00:19:11.031131 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/8816ebee-c735-4b8d-b1ec-65349707b347-ovnkube-script-lib\") pod \"8816ebee-c735-4b8d-b1ec-65349707b347\" (UID: \"8816ebee-c735-4b8d-b1ec-65349707b347\") " Dec 03 00:19:11 crc kubenswrapper[4953]: I1203 00:19:11.031152 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/8816ebee-c735-4b8d-b1ec-65349707b347-ovnkube-config\") pod \"8816ebee-c735-4b8d-b1ec-65349707b347\" (UID: \"8816ebee-c735-4b8d-b1ec-65349707b347\") " Dec 03 00:19:11 crc kubenswrapper[4953]: I1203 00:19:11.031178 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/8816ebee-c735-4b8d-b1ec-65349707b347-systemd-units\") pod \"8816ebee-c735-4b8d-b1ec-65349707b347\" (UID: \"8816ebee-c735-4b8d-b1ec-65349707b347\") " Dec 03 00:19:11 crc kubenswrapper[4953]: I1203 00:19:11.031204 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/8816ebee-c735-4b8d-b1ec-65349707b347-run-systemd\") pod \"8816ebee-c735-4b8d-b1ec-65349707b347\" (UID: \"8816ebee-c735-4b8d-b1ec-65349707b347\") " Dec 03 00:19:11 crc kubenswrapper[4953]: I1203 00:19:11.031226 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/8816ebee-c735-4b8d-b1ec-65349707b347-host-slash\") pod \"8816ebee-c735-4b8d-b1ec-65349707b347\" (UID: \"8816ebee-c735-4b8d-b1ec-65349707b347\") " Dec 03 00:19:11 crc kubenswrapper[4953]: I1203 00:19:11.031247 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/8816ebee-c735-4b8d-b1ec-65349707b347-run-openvswitch\") pod \"8816ebee-c735-4b8d-b1ec-65349707b347\" (UID: \"8816ebee-c735-4b8d-b1ec-65349707b347\") " Dec 03 00:19:11 crc kubenswrapper[4953]: I1203 00:19:11.031258 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8816ebee-c735-4b8d-b1ec-65349707b347-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "8816ebee-c735-4b8d-b1ec-65349707b347" (UID: "8816ebee-c735-4b8d-b1ec-65349707b347"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:19:11 crc kubenswrapper[4953]: I1203 00:19:11.031273 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/8816ebee-c735-4b8d-b1ec-65349707b347-ovn-node-metrics-cert\") pod \"8816ebee-c735-4b8d-b1ec-65349707b347\" (UID: \"8816ebee-c735-4b8d-b1ec-65349707b347\") " Dec 03 00:19:11 crc kubenswrapper[4953]: I1203 00:19:11.031453 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/614357c4-4ad5-4c35-9876-7e4e8ded1c9c-run-openvswitch\") pod \"ovnkube-node-fnl4b\" (UID: \"614357c4-4ad5-4c35-9876-7e4e8ded1c9c\") " pod="openshift-ovn-kubernetes/ovnkube-node-fnl4b" Dec 03 00:19:11 crc kubenswrapper[4953]: I1203 00:19:11.031494 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/614357c4-4ad5-4c35-9876-7e4e8ded1c9c-run-systemd\") pod \"ovnkube-node-fnl4b\" (UID: \"614357c4-4ad5-4c35-9876-7e4e8ded1c9c\") " pod="openshift-ovn-kubernetes/ovnkube-node-fnl4b" Dec 03 00:19:11 crc kubenswrapper[4953]: I1203 00:19:11.031533 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/614357c4-4ad5-4c35-9876-7e4e8ded1c9c-host-slash\") pod \"ovnkube-node-fnl4b\" (UID: \"614357c4-4ad5-4c35-9876-7e4e8ded1c9c\") " pod="openshift-ovn-kubernetes/ovnkube-node-fnl4b" Dec 03 00:19:11 crc kubenswrapper[4953]: I1203 00:19:11.031558 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8816ebee-c735-4b8d-b1ec-65349707b347-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "8816ebee-c735-4b8d-b1ec-65349707b347" (UID: "8816ebee-c735-4b8d-b1ec-65349707b347"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:19:11 crc kubenswrapper[4953]: I1203 00:19:11.031570 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/614357c4-4ad5-4c35-9876-7e4e8ded1c9c-host-kubelet\") pod \"ovnkube-node-fnl4b\" (UID: \"614357c4-4ad5-4c35-9876-7e4e8ded1c9c\") " pod="openshift-ovn-kubernetes/ovnkube-node-fnl4b" Dec 03 00:19:11 crc kubenswrapper[4953]: I1203 00:19:11.031593 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/8816ebee-c735-4b8d-b1ec-65349707b347-host-var-lib-cni-networks-ovn-kubernetes" (OuterVolumeSpecName: "host-var-lib-cni-networks-ovn-kubernetes") pod "8816ebee-c735-4b8d-b1ec-65349707b347" (UID: "8816ebee-c735-4b8d-b1ec-65349707b347"). InnerVolumeSpecName "host-var-lib-cni-networks-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 00:19:11 crc kubenswrapper[4953]: I1203 00:19:11.031598 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/614357c4-4ad5-4c35-9876-7e4e8ded1c9c-env-overrides\") pod \"ovnkube-node-fnl4b\" (UID: \"614357c4-4ad5-4c35-9876-7e4e8ded1c9c\") " pod="openshift-ovn-kubernetes/ovnkube-node-fnl4b" Dec 03 00:19:11 crc kubenswrapper[4953]: I1203 00:19:11.031620 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/8816ebee-c735-4b8d-b1ec-65349707b347-host-cni-netd" (OuterVolumeSpecName: "host-cni-netd") pod "8816ebee-c735-4b8d-b1ec-65349707b347" (UID: "8816ebee-c735-4b8d-b1ec-65349707b347"). InnerVolumeSpecName "host-cni-netd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 00:19:11 crc kubenswrapper[4953]: I1203 00:19:11.031629 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/614357c4-4ad5-4c35-9876-7e4e8ded1c9c-node-log\") pod \"ovnkube-node-fnl4b\" (UID: \"614357c4-4ad5-4c35-9876-7e4e8ded1c9c\") " pod="openshift-ovn-kubernetes/ovnkube-node-fnl4b" Dec 03 00:19:11 crc kubenswrapper[4953]: I1203 00:19:11.031656 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/614357c4-4ad5-4c35-9876-7e4e8ded1c9c-ovnkube-config\") pod \"ovnkube-node-fnl4b\" (UID: \"614357c4-4ad5-4c35-9876-7e4e8ded1c9c\") " pod="openshift-ovn-kubernetes/ovnkube-node-fnl4b" Dec 03 00:19:11 crc kubenswrapper[4953]: I1203 00:19:11.031681 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/614357c4-4ad5-4c35-9876-7e4e8ded1c9c-host-run-ovn-kubernetes\") pod \"ovnkube-node-fnl4b\" (UID: \"614357c4-4ad5-4c35-9876-7e4e8ded1c9c\") " pod="openshift-ovn-kubernetes/ovnkube-node-fnl4b" Dec 03 00:19:11 crc kubenswrapper[4953]: I1203 00:19:11.031704 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/614357c4-4ad5-4c35-9876-7e4e8ded1c9c-run-ovn\") pod \"ovnkube-node-fnl4b\" (UID: \"614357c4-4ad5-4c35-9876-7e4e8ded1c9c\") " pod="openshift-ovn-kubernetes/ovnkube-node-fnl4b" Dec 03 00:19:11 crc kubenswrapper[4953]: I1203 00:19:11.031723 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/614357c4-4ad5-4c35-9876-7e4e8ded1c9c-systemd-units\") pod \"ovnkube-node-fnl4b\" (UID: \"614357c4-4ad5-4c35-9876-7e4e8ded1c9c\") " pod="openshift-ovn-kubernetes/ovnkube-node-fnl4b" Dec 03 00:19:11 crc kubenswrapper[4953]: I1203 00:19:11.031742 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/614357c4-4ad5-4c35-9876-7e4e8ded1c9c-log-socket\") pod \"ovnkube-node-fnl4b\" (UID: \"614357c4-4ad5-4c35-9876-7e4e8ded1c9c\") " pod="openshift-ovn-kubernetes/ovnkube-node-fnl4b" Dec 03 00:19:11 crc kubenswrapper[4953]: I1203 00:19:11.031766 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/614357c4-4ad5-4c35-9876-7e4e8ded1c9c-host-cni-bin\") pod \"ovnkube-node-fnl4b\" (UID: \"614357c4-4ad5-4c35-9876-7e4e8ded1c9c\") " pod="openshift-ovn-kubernetes/ovnkube-node-fnl4b" Dec 03 00:19:11 crc kubenswrapper[4953]: I1203 00:19:11.031786 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/614357c4-4ad5-4c35-9876-7e4e8ded1c9c-host-run-netns\") pod \"ovnkube-node-fnl4b\" (UID: \"614357c4-4ad5-4c35-9876-7e4e8ded1c9c\") " pod="openshift-ovn-kubernetes/ovnkube-node-fnl4b" Dec 03 00:19:11 crc kubenswrapper[4953]: I1203 00:19:11.031810 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/614357c4-4ad5-4c35-9876-7e4e8ded1c9c-ovn-node-metrics-cert\") pod \"ovnkube-node-fnl4b\" (UID: \"614357c4-4ad5-4c35-9876-7e4e8ded1c9c\") " pod="openshift-ovn-kubernetes/ovnkube-node-fnl4b" Dec 03 00:19:11 crc kubenswrapper[4953]: I1203 00:19:11.031830 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/614357c4-4ad5-4c35-9876-7e4e8ded1c9c-ovnkube-script-lib\") pod \"ovnkube-node-fnl4b\" (UID: \"614357c4-4ad5-4c35-9876-7e4e8ded1c9c\") " pod="openshift-ovn-kubernetes/ovnkube-node-fnl4b" Dec 03 00:19:11 crc kubenswrapper[4953]: I1203 00:19:11.031854 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/614357c4-4ad5-4c35-9876-7e4e8ded1c9c-host-cni-netd\") pod \"ovnkube-node-fnl4b\" (UID: \"614357c4-4ad5-4c35-9876-7e4e8ded1c9c\") " pod="openshift-ovn-kubernetes/ovnkube-node-fnl4b" Dec 03 00:19:11 crc kubenswrapper[4953]: I1203 00:19:11.031870 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8816ebee-c735-4b8d-b1ec-65349707b347-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "8816ebee-c735-4b8d-b1ec-65349707b347" (UID: "8816ebee-c735-4b8d-b1ec-65349707b347"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:19:11 crc kubenswrapper[4953]: I1203 00:19:11.031879 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/614357c4-4ad5-4c35-9876-7e4e8ded1c9c-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-fnl4b\" (UID: \"614357c4-4ad5-4c35-9876-7e4e8ded1c9c\") " pod="openshift-ovn-kubernetes/ovnkube-node-fnl4b" Dec 03 00:19:11 crc kubenswrapper[4953]: I1203 00:19:11.031908 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wwdjk\" (UniqueName: \"kubernetes.io/projected/614357c4-4ad5-4c35-9876-7e4e8ded1c9c-kube-api-access-wwdjk\") pod \"ovnkube-node-fnl4b\" (UID: \"614357c4-4ad5-4c35-9876-7e4e8ded1c9c\") " pod="openshift-ovn-kubernetes/ovnkube-node-fnl4b" Dec 03 00:19:11 crc kubenswrapper[4953]: I1203 00:19:11.031938 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/8816ebee-c735-4b8d-b1ec-65349707b347-systemd-units" (OuterVolumeSpecName: "systemd-units") pod "8816ebee-c735-4b8d-b1ec-65349707b347" (UID: "8816ebee-c735-4b8d-b1ec-65349707b347"). InnerVolumeSpecName "systemd-units". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 00:19:11 crc kubenswrapper[4953]: I1203 00:19:11.031939 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/614357c4-4ad5-4c35-9876-7e4e8ded1c9c-var-lib-openvswitch\") pod \"ovnkube-node-fnl4b\" (UID: \"614357c4-4ad5-4c35-9876-7e4e8ded1c9c\") " pod="openshift-ovn-kubernetes/ovnkube-node-fnl4b" Dec 03 00:19:11 crc kubenswrapper[4953]: I1203 00:19:11.032141 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/614357c4-4ad5-4c35-9876-7e4e8ded1c9c-etc-openvswitch\") pod \"ovnkube-node-fnl4b\" (UID: \"614357c4-4ad5-4c35-9876-7e4e8ded1c9c\") " pod="openshift-ovn-kubernetes/ovnkube-node-fnl4b" Dec 03 00:19:11 crc kubenswrapper[4953]: I1203 00:19:11.032396 4953 reconciler_common.go:293] "Volume detached for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/8816ebee-c735-4b8d-b1ec-65349707b347-host-var-lib-cni-networks-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Dec 03 00:19:11 crc kubenswrapper[4953]: I1203 00:19:11.032412 4953 reconciler_common.go:293] "Volume detached for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/8816ebee-c735-4b8d-b1ec-65349707b347-host-cni-netd\") on node \"crc\" DevicePath \"\"" Dec 03 00:19:11 crc kubenswrapper[4953]: I1203 00:19:11.032425 4953 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/8816ebee-c735-4b8d-b1ec-65349707b347-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Dec 03 00:19:11 crc kubenswrapper[4953]: I1203 00:19:11.032439 4953 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/8816ebee-c735-4b8d-b1ec-65349707b347-ovnkube-config\") on node \"crc\" DevicePath \"\"" Dec 03 00:19:11 crc kubenswrapper[4953]: I1203 00:19:11.032451 4953 reconciler_common.go:293] "Volume detached for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/8816ebee-c735-4b8d-b1ec-65349707b347-systemd-units\") on node \"crc\" DevicePath \"\"" Dec 03 00:19:11 crc kubenswrapper[4953]: I1203 00:19:11.032475 4953 reconciler_common.go:293] "Volume detached for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/8816ebee-c735-4b8d-b1ec-65349707b347-host-run-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Dec 03 00:19:11 crc kubenswrapper[4953]: I1203 00:19:11.032493 4953 reconciler_common.go:293] "Volume detached for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/8816ebee-c735-4b8d-b1ec-65349707b347-var-lib-openvswitch\") on node \"crc\" DevicePath \"\"" Dec 03 00:19:11 crc kubenswrapper[4953]: I1203 00:19:11.032507 4953 reconciler_common.go:293] "Volume detached for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/8816ebee-c735-4b8d-b1ec-65349707b347-run-ovn\") on node \"crc\" DevicePath \"\"" Dec 03 00:19:11 crc kubenswrapper[4953]: I1203 00:19:11.032509 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/8816ebee-c735-4b8d-b1ec-65349707b347-host-slash" (OuterVolumeSpecName: "host-slash") pod "8816ebee-c735-4b8d-b1ec-65349707b347" (UID: "8816ebee-c735-4b8d-b1ec-65349707b347"). InnerVolumeSpecName "host-slash". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 00:19:11 crc kubenswrapper[4953]: I1203 00:19:11.032521 4953 reconciler_common.go:293] "Volume detached for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/8816ebee-c735-4b8d-b1ec-65349707b347-host-kubelet\") on node \"crc\" DevicePath \"\"" Dec 03 00:19:11 crc kubenswrapper[4953]: I1203 00:19:11.032568 4953 reconciler_common.go:293] "Volume detached for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/8816ebee-c735-4b8d-b1ec-65349707b347-host-run-netns\") on node \"crc\" DevicePath \"\"" Dec 03 00:19:11 crc kubenswrapper[4953]: I1203 00:19:11.032568 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/8816ebee-c735-4b8d-b1ec-65349707b347-run-openvswitch" (OuterVolumeSpecName: "run-openvswitch") pod "8816ebee-c735-4b8d-b1ec-65349707b347" (UID: "8816ebee-c735-4b8d-b1ec-65349707b347"). InnerVolumeSpecName "run-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 00:19:11 crc kubenswrapper[4953]: I1203 00:19:11.032584 4953 reconciler_common.go:293] "Volume detached for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/8816ebee-c735-4b8d-b1ec-65349707b347-node-log\") on node \"crc\" DevicePath \"\"" Dec 03 00:19:11 crc kubenswrapper[4953]: I1203 00:19:11.032636 4953 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/8816ebee-c735-4b8d-b1ec-65349707b347-env-overrides\") on node \"crc\" DevicePath \"\"" Dec 03 00:19:11 crc kubenswrapper[4953]: I1203 00:19:11.032651 4953 reconciler_common.go:293] "Volume detached for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/8816ebee-c735-4b8d-b1ec-65349707b347-host-cni-bin\") on node \"crc\" DevicePath \"\"" Dec 03 00:19:11 crc kubenswrapper[4953]: I1203 00:19:11.032663 4953 reconciler_common.go:293] "Volume detached for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/8816ebee-c735-4b8d-b1ec-65349707b347-log-socket\") on node \"crc\" DevicePath \"\"" Dec 03 00:19:11 crc kubenswrapper[4953]: I1203 00:19:11.032673 4953 reconciler_common.go:293] "Volume detached for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/8816ebee-c735-4b8d-b1ec-65349707b347-etc-openvswitch\") on node \"crc\" DevicePath \"\"" Dec 03 00:19:11 crc kubenswrapper[4953]: I1203 00:19:11.038864 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8816ebee-c735-4b8d-b1ec-65349707b347-kube-api-access-m2fdf" (OuterVolumeSpecName: "kube-api-access-m2fdf") pod "8816ebee-c735-4b8d-b1ec-65349707b347" (UID: "8816ebee-c735-4b8d-b1ec-65349707b347"). InnerVolumeSpecName "kube-api-access-m2fdf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:19:11 crc kubenswrapper[4953]: I1203 00:19:11.038997 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8816ebee-c735-4b8d-b1ec-65349707b347-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "8816ebee-c735-4b8d-b1ec-65349707b347" (UID: "8816ebee-c735-4b8d-b1ec-65349707b347"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:19:11 crc kubenswrapper[4953]: I1203 00:19:11.049570 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/8816ebee-c735-4b8d-b1ec-65349707b347-run-systemd" (OuterVolumeSpecName: "run-systemd") pod "8816ebee-c735-4b8d-b1ec-65349707b347" (UID: "8816ebee-c735-4b8d-b1ec-65349707b347"). InnerVolumeSpecName "run-systemd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 00:19:11 crc kubenswrapper[4953]: I1203 00:19:11.134136 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/614357c4-4ad5-4c35-9876-7e4e8ded1c9c-ovn-node-metrics-cert\") pod \"ovnkube-node-fnl4b\" (UID: \"614357c4-4ad5-4c35-9876-7e4e8ded1c9c\") " pod="openshift-ovn-kubernetes/ovnkube-node-fnl4b" Dec 03 00:19:11 crc kubenswrapper[4953]: I1203 00:19:11.134591 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/614357c4-4ad5-4c35-9876-7e4e8ded1c9c-ovnkube-script-lib\") pod \"ovnkube-node-fnl4b\" (UID: \"614357c4-4ad5-4c35-9876-7e4e8ded1c9c\") " pod="openshift-ovn-kubernetes/ovnkube-node-fnl4b" Dec 03 00:19:11 crc kubenswrapper[4953]: I1203 00:19:11.134613 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/614357c4-4ad5-4c35-9876-7e4e8ded1c9c-host-cni-netd\") pod \"ovnkube-node-fnl4b\" (UID: \"614357c4-4ad5-4c35-9876-7e4e8ded1c9c\") " pod="openshift-ovn-kubernetes/ovnkube-node-fnl4b" Dec 03 00:19:11 crc kubenswrapper[4953]: I1203 00:19:11.134631 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/614357c4-4ad5-4c35-9876-7e4e8ded1c9c-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-fnl4b\" (UID: \"614357c4-4ad5-4c35-9876-7e4e8ded1c9c\") " pod="openshift-ovn-kubernetes/ovnkube-node-fnl4b" Dec 03 00:19:11 crc kubenswrapper[4953]: I1203 00:19:11.134665 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wwdjk\" (UniqueName: \"kubernetes.io/projected/614357c4-4ad5-4c35-9876-7e4e8ded1c9c-kube-api-access-wwdjk\") pod \"ovnkube-node-fnl4b\" (UID: \"614357c4-4ad5-4c35-9876-7e4e8ded1c9c\") " pod="openshift-ovn-kubernetes/ovnkube-node-fnl4b" Dec 03 00:19:11 crc kubenswrapper[4953]: I1203 00:19:11.134693 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/614357c4-4ad5-4c35-9876-7e4e8ded1c9c-var-lib-openvswitch\") pod \"ovnkube-node-fnl4b\" (UID: \"614357c4-4ad5-4c35-9876-7e4e8ded1c9c\") " pod="openshift-ovn-kubernetes/ovnkube-node-fnl4b" Dec 03 00:19:11 crc kubenswrapper[4953]: I1203 00:19:11.134725 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/614357c4-4ad5-4c35-9876-7e4e8ded1c9c-etc-openvswitch\") pod \"ovnkube-node-fnl4b\" (UID: \"614357c4-4ad5-4c35-9876-7e4e8ded1c9c\") " pod="openshift-ovn-kubernetes/ovnkube-node-fnl4b" Dec 03 00:19:11 crc kubenswrapper[4953]: I1203 00:19:11.134758 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/614357c4-4ad5-4c35-9876-7e4e8ded1c9c-run-openvswitch\") pod \"ovnkube-node-fnl4b\" (UID: \"614357c4-4ad5-4c35-9876-7e4e8ded1c9c\") " pod="openshift-ovn-kubernetes/ovnkube-node-fnl4b" Dec 03 00:19:11 crc kubenswrapper[4953]: I1203 00:19:11.134762 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/614357c4-4ad5-4c35-9876-7e4e8ded1c9c-host-cni-netd\") pod \"ovnkube-node-fnl4b\" (UID: \"614357c4-4ad5-4c35-9876-7e4e8ded1c9c\") " pod="openshift-ovn-kubernetes/ovnkube-node-fnl4b" Dec 03 00:19:11 crc kubenswrapper[4953]: I1203 00:19:11.134863 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/614357c4-4ad5-4c35-9876-7e4e8ded1c9c-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-fnl4b\" (UID: \"614357c4-4ad5-4c35-9876-7e4e8ded1c9c\") " pod="openshift-ovn-kubernetes/ovnkube-node-fnl4b" Dec 03 00:19:11 crc kubenswrapper[4953]: I1203 00:19:11.134790 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/614357c4-4ad5-4c35-9876-7e4e8ded1c9c-run-systemd\") pod \"ovnkube-node-fnl4b\" (UID: \"614357c4-4ad5-4c35-9876-7e4e8ded1c9c\") " pod="openshift-ovn-kubernetes/ovnkube-node-fnl4b" Dec 03 00:19:11 crc kubenswrapper[4953]: I1203 00:19:11.135379 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/614357c4-4ad5-4c35-9876-7e4e8ded1c9c-etc-openvswitch\") pod \"ovnkube-node-fnl4b\" (UID: \"614357c4-4ad5-4c35-9876-7e4e8ded1c9c\") " pod="openshift-ovn-kubernetes/ovnkube-node-fnl4b" Dec 03 00:19:11 crc kubenswrapper[4953]: I1203 00:19:11.135425 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/614357c4-4ad5-4c35-9876-7e4e8ded1c9c-run-openvswitch\") pod \"ovnkube-node-fnl4b\" (UID: \"614357c4-4ad5-4c35-9876-7e4e8ded1c9c\") " pod="openshift-ovn-kubernetes/ovnkube-node-fnl4b" Dec 03 00:19:11 crc kubenswrapper[4953]: I1203 00:19:11.135420 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/614357c4-4ad5-4c35-9876-7e4e8ded1c9c-var-lib-openvswitch\") pod \"ovnkube-node-fnl4b\" (UID: \"614357c4-4ad5-4c35-9876-7e4e8ded1c9c\") " pod="openshift-ovn-kubernetes/ovnkube-node-fnl4b" Dec 03 00:19:11 crc kubenswrapper[4953]: I1203 00:19:11.135470 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/614357c4-4ad5-4c35-9876-7e4e8ded1c9c-run-systemd\") pod \"ovnkube-node-fnl4b\" (UID: \"614357c4-4ad5-4c35-9876-7e4e8ded1c9c\") " pod="openshift-ovn-kubernetes/ovnkube-node-fnl4b" Dec 03 00:19:11 crc kubenswrapper[4953]: I1203 00:19:11.135855 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/614357c4-4ad5-4c35-9876-7e4e8ded1c9c-ovnkube-script-lib\") pod \"ovnkube-node-fnl4b\" (UID: \"614357c4-4ad5-4c35-9876-7e4e8ded1c9c\") " pod="openshift-ovn-kubernetes/ovnkube-node-fnl4b" Dec 03 00:19:11 crc kubenswrapper[4953]: I1203 00:19:11.136085 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/614357c4-4ad5-4c35-9876-7e4e8ded1c9c-host-slash\") pod \"ovnkube-node-fnl4b\" (UID: \"614357c4-4ad5-4c35-9876-7e4e8ded1c9c\") " pod="openshift-ovn-kubernetes/ovnkube-node-fnl4b" Dec 03 00:19:11 crc kubenswrapper[4953]: I1203 00:19:11.136170 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/614357c4-4ad5-4c35-9876-7e4e8ded1c9c-host-slash\") pod \"ovnkube-node-fnl4b\" (UID: \"614357c4-4ad5-4c35-9876-7e4e8ded1c9c\") " pod="openshift-ovn-kubernetes/ovnkube-node-fnl4b" Dec 03 00:19:11 crc kubenswrapper[4953]: I1203 00:19:11.136216 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/614357c4-4ad5-4c35-9876-7e4e8ded1c9c-host-kubelet\") pod \"ovnkube-node-fnl4b\" (UID: \"614357c4-4ad5-4c35-9876-7e4e8ded1c9c\") " pod="openshift-ovn-kubernetes/ovnkube-node-fnl4b" Dec 03 00:19:11 crc kubenswrapper[4953]: I1203 00:19:11.136325 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/614357c4-4ad5-4c35-9876-7e4e8ded1c9c-host-kubelet\") pod \"ovnkube-node-fnl4b\" (UID: \"614357c4-4ad5-4c35-9876-7e4e8ded1c9c\") " pod="openshift-ovn-kubernetes/ovnkube-node-fnl4b" Dec 03 00:19:11 crc kubenswrapper[4953]: I1203 00:19:11.136333 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/614357c4-4ad5-4c35-9876-7e4e8ded1c9c-env-overrides\") pod \"ovnkube-node-fnl4b\" (UID: \"614357c4-4ad5-4c35-9876-7e4e8ded1c9c\") " pod="openshift-ovn-kubernetes/ovnkube-node-fnl4b" Dec 03 00:19:11 crc kubenswrapper[4953]: I1203 00:19:11.136389 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/614357c4-4ad5-4c35-9876-7e4e8ded1c9c-node-log\") pod \"ovnkube-node-fnl4b\" (UID: \"614357c4-4ad5-4c35-9876-7e4e8ded1c9c\") " pod="openshift-ovn-kubernetes/ovnkube-node-fnl4b" Dec 03 00:19:11 crc kubenswrapper[4953]: I1203 00:19:11.136421 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/614357c4-4ad5-4c35-9876-7e4e8ded1c9c-ovnkube-config\") pod \"ovnkube-node-fnl4b\" (UID: \"614357c4-4ad5-4c35-9876-7e4e8ded1c9c\") " pod="openshift-ovn-kubernetes/ovnkube-node-fnl4b" Dec 03 00:19:11 crc kubenswrapper[4953]: I1203 00:19:11.136450 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/614357c4-4ad5-4c35-9876-7e4e8ded1c9c-host-run-ovn-kubernetes\") pod \"ovnkube-node-fnl4b\" (UID: \"614357c4-4ad5-4c35-9876-7e4e8ded1c9c\") " pod="openshift-ovn-kubernetes/ovnkube-node-fnl4b" Dec 03 00:19:11 crc kubenswrapper[4953]: I1203 00:19:11.136482 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/614357c4-4ad5-4c35-9876-7e4e8ded1c9c-run-ovn\") pod \"ovnkube-node-fnl4b\" (UID: \"614357c4-4ad5-4c35-9876-7e4e8ded1c9c\") " pod="openshift-ovn-kubernetes/ovnkube-node-fnl4b" Dec 03 00:19:11 crc kubenswrapper[4953]: I1203 00:19:11.136514 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/614357c4-4ad5-4c35-9876-7e4e8ded1c9c-systemd-units\") pod \"ovnkube-node-fnl4b\" (UID: \"614357c4-4ad5-4c35-9876-7e4e8ded1c9c\") " pod="openshift-ovn-kubernetes/ovnkube-node-fnl4b" Dec 03 00:19:11 crc kubenswrapper[4953]: I1203 00:19:11.136542 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/614357c4-4ad5-4c35-9876-7e4e8ded1c9c-log-socket\") pod \"ovnkube-node-fnl4b\" (UID: \"614357c4-4ad5-4c35-9876-7e4e8ded1c9c\") " pod="openshift-ovn-kubernetes/ovnkube-node-fnl4b" Dec 03 00:19:11 crc kubenswrapper[4953]: I1203 00:19:11.136585 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/614357c4-4ad5-4c35-9876-7e4e8ded1c9c-host-cni-bin\") pod \"ovnkube-node-fnl4b\" (UID: \"614357c4-4ad5-4c35-9876-7e4e8ded1c9c\") " pod="openshift-ovn-kubernetes/ovnkube-node-fnl4b" Dec 03 00:19:11 crc kubenswrapper[4953]: I1203 00:19:11.136611 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/614357c4-4ad5-4c35-9876-7e4e8ded1c9c-host-run-netns\") pod \"ovnkube-node-fnl4b\" (UID: \"614357c4-4ad5-4c35-9876-7e4e8ded1c9c\") " pod="openshift-ovn-kubernetes/ovnkube-node-fnl4b" Dec 03 00:19:11 crc kubenswrapper[4953]: I1203 00:19:11.136660 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/614357c4-4ad5-4c35-9876-7e4e8ded1c9c-host-run-ovn-kubernetes\") pod \"ovnkube-node-fnl4b\" (UID: \"614357c4-4ad5-4c35-9876-7e4e8ded1c9c\") " pod="openshift-ovn-kubernetes/ovnkube-node-fnl4b" Dec 03 00:19:11 crc kubenswrapper[4953]: I1203 00:19:11.136612 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/614357c4-4ad5-4c35-9876-7e4e8ded1c9c-run-ovn\") pod \"ovnkube-node-fnl4b\" (UID: \"614357c4-4ad5-4c35-9876-7e4e8ded1c9c\") " pod="openshift-ovn-kubernetes/ovnkube-node-fnl4b" Dec 03 00:19:11 crc kubenswrapper[4953]: I1203 00:19:11.136738 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/614357c4-4ad5-4c35-9876-7e4e8ded1c9c-log-socket\") pod \"ovnkube-node-fnl4b\" (UID: \"614357c4-4ad5-4c35-9876-7e4e8ded1c9c\") " pod="openshift-ovn-kubernetes/ovnkube-node-fnl4b" Dec 03 00:19:11 crc kubenswrapper[4953]: I1203 00:19:11.136750 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/614357c4-4ad5-4c35-9876-7e4e8ded1c9c-host-run-netns\") pod \"ovnkube-node-fnl4b\" (UID: \"614357c4-4ad5-4c35-9876-7e4e8ded1c9c\") " pod="openshift-ovn-kubernetes/ovnkube-node-fnl4b" Dec 03 00:19:11 crc kubenswrapper[4953]: I1203 00:19:11.136758 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/614357c4-4ad5-4c35-9876-7e4e8ded1c9c-systemd-units\") pod \"ovnkube-node-fnl4b\" (UID: \"614357c4-4ad5-4c35-9876-7e4e8ded1c9c\") " pod="openshift-ovn-kubernetes/ovnkube-node-fnl4b" Dec 03 00:19:11 crc kubenswrapper[4953]: I1203 00:19:11.136782 4953 reconciler_common.go:293] "Volume detached for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/8816ebee-c735-4b8d-b1ec-65349707b347-run-systemd\") on node \"crc\" DevicePath \"\"" Dec 03 00:19:11 crc kubenswrapper[4953]: I1203 00:19:11.136803 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/614357c4-4ad5-4c35-9876-7e4e8ded1c9c-node-log\") pod \"ovnkube-node-fnl4b\" (UID: \"614357c4-4ad5-4c35-9876-7e4e8ded1c9c\") " pod="openshift-ovn-kubernetes/ovnkube-node-fnl4b" Dec 03 00:19:11 crc kubenswrapper[4953]: I1203 00:19:11.136823 4953 reconciler_common.go:293] "Volume detached for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/8816ebee-c735-4b8d-b1ec-65349707b347-host-slash\") on node \"crc\" DevicePath \"\"" Dec 03 00:19:11 crc kubenswrapper[4953]: I1203 00:19:11.136841 4953 reconciler_common.go:293] "Volume detached for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/8816ebee-c735-4b8d-b1ec-65349707b347-run-openvswitch\") on node \"crc\" DevicePath \"\"" Dec 03 00:19:11 crc kubenswrapper[4953]: I1203 00:19:11.136853 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/614357c4-4ad5-4c35-9876-7e4e8ded1c9c-host-cni-bin\") pod \"ovnkube-node-fnl4b\" (UID: \"614357c4-4ad5-4c35-9876-7e4e8ded1c9c\") " pod="openshift-ovn-kubernetes/ovnkube-node-fnl4b" Dec 03 00:19:11 crc kubenswrapper[4953]: I1203 00:19:11.136856 4953 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/8816ebee-c735-4b8d-b1ec-65349707b347-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Dec 03 00:19:11 crc kubenswrapper[4953]: I1203 00:19:11.136925 4953 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m2fdf\" (UniqueName: \"kubernetes.io/projected/8816ebee-c735-4b8d-b1ec-65349707b347-kube-api-access-m2fdf\") on node \"crc\" DevicePath \"\"" Dec 03 00:19:11 crc kubenswrapper[4953]: I1203 00:19:11.137698 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/614357c4-4ad5-4c35-9876-7e4e8ded1c9c-ovnkube-config\") pod \"ovnkube-node-fnl4b\" (UID: \"614357c4-4ad5-4c35-9876-7e4e8ded1c9c\") " pod="openshift-ovn-kubernetes/ovnkube-node-fnl4b" Dec 03 00:19:11 crc kubenswrapper[4953]: I1203 00:19:11.137697 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/614357c4-4ad5-4c35-9876-7e4e8ded1c9c-env-overrides\") pod \"ovnkube-node-fnl4b\" (UID: \"614357c4-4ad5-4c35-9876-7e4e8ded1c9c\") " pod="openshift-ovn-kubernetes/ovnkube-node-fnl4b" Dec 03 00:19:11 crc kubenswrapper[4953]: I1203 00:19:11.138886 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/614357c4-4ad5-4c35-9876-7e4e8ded1c9c-ovn-node-metrics-cert\") pod \"ovnkube-node-fnl4b\" (UID: \"614357c4-4ad5-4c35-9876-7e4e8ded1c9c\") " pod="openshift-ovn-kubernetes/ovnkube-node-fnl4b" Dec 03 00:19:11 crc kubenswrapper[4953]: I1203 00:19:11.162297 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wwdjk\" (UniqueName: \"kubernetes.io/projected/614357c4-4ad5-4c35-9876-7e4e8ded1c9c-kube-api-access-wwdjk\") pod \"ovnkube-node-fnl4b\" (UID: \"614357c4-4ad5-4c35-9876-7e4e8ded1c9c\") " pod="openshift-ovn-kubernetes/ovnkube-node-fnl4b" Dec 03 00:19:11 crc kubenswrapper[4953]: I1203 00:19:11.237068 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-fnl4b" Dec 03 00:19:11 crc kubenswrapper[4953]: W1203 00:19:11.257945 4953 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod614357c4_4ad5_4c35_9876_7e4e8ded1c9c.slice/crio-fc33cd02817003207736f41f45702e6ff1277a84015f3cb5c235e991c866218f WatchSource:0}: Error finding container fc33cd02817003207736f41f45702e6ff1277a84015f3cb5c235e991c866218f: Status 404 returned error can't find the container with id fc33cd02817003207736f41f45702e6ff1277a84015f3cb5c235e991c866218f Dec 03 00:19:11 crc kubenswrapper[4953]: I1203 00:19:11.571356 4953 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-txffx_8816ebee-c735-4b8d-b1ec-65349707b347/ovnkube-controller/3.log" Dec 03 00:19:11 crc kubenswrapper[4953]: I1203 00:19:11.576198 4953 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-txffx_8816ebee-c735-4b8d-b1ec-65349707b347/ovn-acl-logging/0.log" Dec 03 00:19:11 crc kubenswrapper[4953]: I1203 00:19:11.576821 4953 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-txffx_8816ebee-c735-4b8d-b1ec-65349707b347/ovn-controller/0.log" Dec 03 00:19:11 crc kubenswrapper[4953]: I1203 00:19:11.577268 4953 generic.go:334] "Generic (PLEG): container finished" podID="8816ebee-c735-4b8d-b1ec-65349707b347" containerID="bcde8ca2a20e2afee0f6cb7164e6068aa82f05c0c978ea3cc4c23618fcb57bac" exitCode=0 Dec 03 00:19:11 crc kubenswrapper[4953]: I1203 00:19:11.577346 4953 generic.go:334] "Generic (PLEG): container finished" podID="8816ebee-c735-4b8d-b1ec-65349707b347" containerID="7deadc8e135ff1bc7c5a2f093e489bb5b4b6f166bc874c6acc2a841c5397f11e" exitCode=0 Dec 03 00:19:11 crc kubenswrapper[4953]: I1203 00:19:11.577411 4953 generic.go:334] "Generic (PLEG): container finished" podID="8816ebee-c735-4b8d-b1ec-65349707b347" containerID="a61b951a9cbab2c48c0f2df8351f97d7ff71bf3674f641fe5eac30155232b224" exitCode=0 Dec 03 00:19:11 crc kubenswrapper[4953]: I1203 00:19:11.577363 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-txffx" event={"ID":"8816ebee-c735-4b8d-b1ec-65349707b347","Type":"ContainerDied","Data":"bcde8ca2a20e2afee0f6cb7164e6068aa82f05c0c978ea3cc4c23618fcb57bac"} Dec 03 00:19:11 crc kubenswrapper[4953]: I1203 00:19:11.577531 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-txffx" event={"ID":"8816ebee-c735-4b8d-b1ec-65349707b347","Type":"ContainerDied","Data":"7deadc8e135ff1bc7c5a2f093e489bb5b4b6f166bc874c6acc2a841c5397f11e"} Dec 03 00:19:11 crc kubenswrapper[4953]: I1203 00:19:11.577560 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-txffx" event={"ID":"8816ebee-c735-4b8d-b1ec-65349707b347","Type":"ContainerDied","Data":"a61b951a9cbab2c48c0f2df8351f97d7ff71bf3674f641fe5eac30155232b224"} Dec 03 00:19:11 crc kubenswrapper[4953]: I1203 00:19:11.577578 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-txffx" event={"ID":"8816ebee-c735-4b8d-b1ec-65349707b347","Type":"ContainerDied","Data":"fd5f428de229749485267e81d34e31b708728b52474032e88460e54747f21e11"} Dec 03 00:19:11 crc kubenswrapper[4953]: I1203 00:19:11.577478 4953 generic.go:334] "Generic (PLEG): container finished" podID="8816ebee-c735-4b8d-b1ec-65349707b347" containerID="fd5f428de229749485267e81d34e31b708728b52474032e88460e54747f21e11" exitCode=0 Dec 03 00:19:11 crc kubenswrapper[4953]: I1203 00:19:11.577602 4953 scope.go:117] "RemoveContainer" containerID="bcde8ca2a20e2afee0f6cb7164e6068aa82f05c0c978ea3cc4c23618fcb57bac" Dec 03 00:19:11 crc kubenswrapper[4953]: I1203 00:19:11.577643 4953 generic.go:334] "Generic (PLEG): container finished" podID="8816ebee-c735-4b8d-b1ec-65349707b347" containerID="e0f6f645a58993f052a6f365bb808f62d99f7fea74886c5cc022b2d86112321a" exitCode=0 Dec 03 00:19:11 crc kubenswrapper[4953]: I1203 00:19:11.577764 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-txffx" event={"ID":"8816ebee-c735-4b8d-b1ec-65349707b347","Type":"ContainerDied","Data":"e0f6f645a58993f052a6f365bb808f62d99f7fea74886c5cc022b2d86112321a"} Dec 03 00:19:11 crc kubenswrapper[4953]: I1203 00:19:11.577791 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-txffx" event={"ID":"8816ebee-c735-4b8d-b1ec-65349707b347","Type":"ContainerDied","Data":"d4b85224419a4075acb340e4b11095f2adb5c316e285afa4e19295c2181c0d24"} Dec 03 00:19:11 crc kubenswrapper[4953]: I1203 00:19:11.577808 4953 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"0a4b625eda9b9daff8aa7fe3c5b94e417bcefe08fc6a65b4594f821cfdca7bfd"} Dec 03 00:19:11 crc kubenswrapper[4953]: I1203 00:19:11.577823 4953 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"7deadc8e135ff1bc7c5a2f093e489bb5b4b6f166bc874c6acc2a841c5397f11e"} Dec 03 00:19:11 crc kubenswrapper[4953]: I1203 00:19:11.577831 4953 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"a61b951a9cbab2c48c0f2df8351f97d7ff71bf3674f641fe5eac30155232b224"} Dec 03 00:19:11 crc kubenswrapper[4953]: I1203 00:19:11.577840 4953 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"fd5f428de229749485267e81d34e31b708728b52474032e88460e54747f21e11"} Dec 03 00:19:11 crc kubenswrapper[4953]: I1203 00:19:11.577848 4953 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"e0f6f645a58993f052a6f365bb808f62d99f7fea74886c5cc022b2d86112321a"} Dec 03 00:19:11 crc kubenswrapper[4953]: I1203 00:19:11.577857 4953 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"d4b85224419a4075acb340e4b11095f2adb5c316e285afa4e19295c2181c0d24"} Dec 03 00:19:11 crc kubenswrapper[4953]: I1203 00:19:11.577866 4953 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"00a2ea4a9bb870fce30e7fa845f787f184e7a19e42c95511655ea8c5f84a26d1"} Dec 03 00:19:11 crc kubenswrapper[4953]: I1203 00:19:11.577874 4953 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"dc0a5d4cc9d614d8492626a1b114bdc52f510eb5965d842c02c083f08d4849a7"} Dec 03 00:19:11 crc kubenswrapper[4953]: I1203 00:19:11.577888 4953 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"af4c7764fca87144d01f3738ea010308744ccf50c54b7c471ccc213112de00d5"} Dec 03 00:19:11 crc kubenswrapper[4953]: I1203 00:19:11.577685 4953 generic.go:334] "Generic (PLEG): container finished" podID="8816ebee-c735-4b8d-b1ec-65349707b347" containerID="d4b85224419a4075acb340e4b11095f2adb5c316e285afa4e19295c2181c0d24" exitCode=0 Dec 03 00:19:11 crc kubenswrapper[4953]: I1203 00:19:11.577920 4953 generic.go:334] "Generic (PLEG): container finished" podID="8816ebee-c735-4b8d-b1ec-65349707b347" containerID="00a2ea4a9bb870fce30e7fa845f787f184e7a19e42c95511655ea8c5f84a26d1" exitCode=143 Dec 03 00:19:11 crc kubenswrapper[4953]: I1203 00:19:11.577935 4953 generic.go:334] "Generic (PLEG): container finished" podID="8816ebee-c735-4b8d-b1ec-65349707b347" containerID="dc0a5d4cc9d614d8492626a1b114bdc52f510eb5965d842c02c083f08d4849a7" exitCode=143 Dec 03 00:19:11 crc kubenswrapper[4953]: I1203 00:19:11.578020 4953 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-txffx" Dec 03 00:19:11 crc kubenswrapper[4953]: I1203 00:19:11.578224 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-txffx" event={"ID":"8816ebee-c735-4b8d-b1ec-65349707b347","Type":"ContainerDied","Data":"00a2ea4a9bb870fce30e7fa845f787f184e7a19e42c95511655ea8c5f84a26d1"} Dec 03 00:19:11 crc kubenswrapper[4953]: I1203 00:19:11.578269 4953 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"bcde8ca2a20e2afee0f6cb7164e6068aa82f05c0c978ea3cc4c23618fcb57bac"} Dec 03 00:19:11 crc kubenswrapper[4953]: I1203 00:19:11.578284 4953 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"0a4b625eda9b9daff8aa7fe3c5b94e417bcefe08fc6a65b4594f821cfdca7bfd"} Dec 03 00:19:11 crc kubenswrapper[4953]: I1203 00:19:11.578298 4953 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"7deadc8e135ff1bc7c5a2f093e489bb5b4b6f166bc874c6acc2a841c5397f11e"} Dec 03 00:19:11 crc kubenswrapper[4953]: I1203 00:19:11.578306 4953 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"a61b951a9cbab2c48c0f2df8351f97d7ff71bf3674f641fe5eac30155232b224"} Dec 03 00:19:11 crc kubenswrapper[4953]: I1203 00:19:11.578313 4953 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"fd5f428de229749485267e81d34e31b708728b52474032e88460e54747f21e11"} Dec 03 00:19:11 crc kubenswrapper[4953]: I1203 00:19:11.578320 4953 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"e0f6f645a58993f052a6f365bb808f62d99f7fea74886c5cc022b2d86112321a"} Dec 03 00:19:11 crc kubenswrapper[4953]: I1203 00:19:11.578327 4953 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"d4b85224419a4075acb340e4b11095f2adb5c316e285afa4e19295c2181c0d24"} Dec 03 00:19:11 crc kubenswrapper[4953]: I1203 00:19:11.578337 4953 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"00a2ea4a9bb870fce30e7fa845f787f184e7a19e42c95511655ea8c5f84a26d1"} Dec 03 00:19:11 crc kubenswrapper[4953]: I1203 00:19:11.578343 4953 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"dc0a5d4cc9d614d8492626a1b114bdc52f510eb5965d842c02c083f08d4849a7"} Dec 03 00:19:11 crc kubenswrapper[4953]: I1203 00:19:11.578350 4953 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"af4c7764fca87144d01f3738ea010308744ccf50c54b7c471ccc213112de00d5"} Dec 03 00:19:11 crc kubenswrapper[4953]: I1203 00:19:11.578366 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-txffx" event={"ID":"8816ebee-c735-4b8d-b1ec-65349707b347","Type":"ContainerDied","Data":"dc0a5d4cc9d614d8492626a1b114bdc52f510eb5965d842c02c083f08d4849a7"} Dec 03 00:19:11 crc kubenswrapper[4953]: I1203 00:19:11.578387 4953 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"bcde8ca2a20e2afee0f6cb7164e6068aa82f05c0c978ea3cc4c23618fcb57bac"} Dec 03 00:19:11 crc kubenswrapper[4953]: I1203 00:19:11.578396 4953 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"0a4b625eda9b9daff8aa7fe3c5b94e417bcefe08fc6a65b4594f821cfdca7bfd"} Dec 03 00:19:11 crc kubenswrapper[4953]: I1203 00:19:11.578403 4953 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"7deadc8e135ff1bc7c5a2f093e489bb5b4b6f166bc874c6acc2a841c5397f11e"} Dec 03 00:19:11 crc kubenswrapper[4953]: I1203 00:19:11.578408 4953 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"a61b951a9cbab2c48c0f2df8351f97d7ff71bf3674f641fe5eac30155232b224"} Dec 03 00:19:11 crc kubenswrapper[4953]: I1203 00:19:11.578413 4953 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"fd5f428de229749485267e81d34e31b708728b52474032e88460e54747f21e11"} Dec 03 00:19:11 crc kubenswrapper[4953]: I1203 00:19:11.578419 4953 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"e0f6f645a58993f052a6f365bb808f62d99f7fea74886c5cc022b2d86112321a"} Dec 03 00:19:11 crc kubenswrapper[4953]: I1203 00:19:11.578427 4953 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"d4b85224419a4075acb340e4b11095f2adb5c316e285afa4e19295c2181c0d24"} Dec 03 00:19:11 crc kubenswrapper[4953]: I1203 00:19:11.578433 4953 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"00a2ea4a9bb870fce30e7fa845f787f184e7a19e42c95511655ea8c5f84a26d1"} Dec 03 00:19:11 crc kubenswrapper[4953]: I1203 00:19:11.578439 4953 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"dc0a5d4cc9d614d8492626a1b114bdc52f510eb5965d842c02c083f08d4849a7"} Dec 03 00:19:11 crc kubenswrapper[4953]: I1203 00:19:11.578445 4953 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"af4c7764fca87144d01f3738ea010308744ccf50c54b7c471ccc213112de00d5"} Dec 03 00:19:11 crc kubenswrapper[4953]: I1203 00:19:11.578453 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-txffx" event={"ID":"8816ebee-c735-4b8d-b1ec-65349707b347","Type":"ContainerDied","Data":"3780938f7eb5daa186db861efd64edfb729e40aaabab9ded50425004f8f16d42"} Dec 03 00:19:11 crc kubenswrapper[4953]: I1203 00:19:11.578461 4953 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"bcde8ca2a20e2afee0f6cb7164e6068aa82f05c0c978ea3cc4c23618fcb57bac"} Dec 03 00:19:11 crc kubenswrapper[4953]: I1203 00:19:11.578468 4953 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"0a4b625eda9b9daff8aa7fe3c5b94e417bcefe08fc6a65b4594f821cfdca7bfd"} Dec 03 00:19:11 crc kubenswrapper[4953]: I1203 00:19:11.578473 4953 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"7deadc8e135ff1bc7c5a2f093e489bb5b4b6f166bc874c6acc2a841c5397f11e"} Dec 03 00:19:11 crc kubenswrapper[4953]: I1203 00:19:11.578478 4953 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"a61b951a9cbab2c48c0f2df8351f97d7ff71bf3674f641fe5eac30155232b224"} Dec 03 00:19:11 crc kubenswrapper[4953]: I1203 00:19:11.578484 4953 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"fd5f428de229749485267e81d34e31b708728b52474032e88460e54747f21e11"} Dec 03 00:19:11 crc kubenswrapper[4953]: I1203 00:19:11.578489 4953 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"e0f6f645a58993f052a6f365bb808f62d99f7fea74886c5cc022b2d86112321a"} Dec 03 00:19:11 crc kubenswrapper[4953]: I1203 00:19:11.578494 4953 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"d4b85224419a4075acb340e4b11095f2adb5c316e285afa4e19295c2181c0d24"} Dec 03 00:19:11 crc kubenswrapper[4953]: I1203 00:19:11.578500 4953 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"00a2ea4a9bb870fce30e7fa845f787f184e7a19e42c95511655ea8c5f84a26d1"} Dec 03 00:19:11 crc kubenswrapper[4953]: I1203 00:19:11.578505 4953 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"dc0a5d4cc9d614d8492626a1b114bdc52f510eb5965d842c02c083f08d4849a7"} Dec 03 00:19:11 crc kubenswrapper[4953]: I1203 00:19:11.578510 4953 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"af4c7764fca87144d01f3738ea010308744ccf50c54b7c471ccc213112de00d5"} Dec 03 00:19:11 crc kubenswrapper[4953]: I1203 00:19:11.583554 4953 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-675qk_43a4bad8-71f4-441a-8302-aeb526a986af/kube-multus/2.log" Dec 03 00:19:11 crc kubenswrapper[4953]: I1203 00:19:11.583692 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-675qk" event={"ID":"43a4bad8-71f4-441a-8302-aeb526a986af","Type":"ContainerStarted","Data":"c349af53feb62a5f26511614c2678f012aa22d0250851fa32f796a2f62f2155a"} Dec 03 00:19:11 crc kubenswrapper[4953]: I1203 00:19:11.585729 4953 generic.go:334] "Generic (PLEG): container finished" podID="614357c4-4ad5-4c35-9876-7e4e8ded1c9c" containerID="cb25f20ebcbe271db31fc13dade09a7cadad925114c1f2235778d8f71e9da254" exitCode=0 Dec 03 00:19:11 crc kubenswrapper[4953]: I1203 00:19:11.585769 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fnl4b" event={"ID":"614357c4-4ad5-4c35-9876-7e4e8ded1c9c","Type":"ContainerDied","Data":"cb25f20ebcbe271db31fc13dade09a7cadad925114c1f2235778d8f71e9da254"} Dec 03 00:19:11 crc kubenswrapper[4953]: I1203 00:19:11.585818 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fnl4b" event={"ID":"614357c4-4ad5-4c35-9876-7e4e8ded1c9c","Type":"ContainerStarted","Data":"fc33cd02817003207736f41f45702e6ff1277a84015f3cb5c235e991c866218f"} Dec 03 00:19:11 crc kubenswrapper[4953]: I1203 00:19:11.607938 4953 scope.go:117] "RemoveContainer" containerID="0a4b625eda9b9daff8aa7fe3c5b94e417bcefe08fc6a65b4594f821cfdca7bfd" Dec 03 00:19:11 crc kubenswrapper[4953]: I1203 00:19:11.610369 4953 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-txffx"] Dec 03 00:19:11 crc kubenswrapper[4953]: I1203 00:19:11.614851 4953 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-txffx"] Dec 03 00:19:11 crc kubenswrapper[4953]: I1203 00:19:11.679638 4953 scope.go:117] "RemoveContainer" containerID="7deadc8e135ff1bc7c5a2f093e489bb5b4b6f166bc874c6acc2a841c5397f11e" Dec 03 00:19:11 crc kubenswrapper[4953]: I1203 00:19:11.699789 4953 scope.go:117] "RemoveContainer" containerID="a61b951a9cbab2c48c0f2df8351f97d7ff71bf3674f641fe5eac30155232b224" Dec 03 00:19:11 crc kubenswrapper[4953]: I1203 00:19:11.729312 4953 scope.go:117] "RemoveContainer" containerID="fd5f428de229749485267e81d34e31b708728b52474032e88460e54747f21e11" Dec 03 00:19:11 crc kubenswrapper[4953]: I1203 00:19:11.748297 4953 scope.go:117] "RemoveContainer" containerID="e0f6f645a58993f052a6f365bb808f62d99f7fea74886c5cc022b2d86112321a" Dec 03 00:19:11 crc kubenswrapper[4953]: I1203 00:19:11.766083 4953 scope.go:117] "RemoveContainer" containerID="d4b85224419a4075acb340e4b11095f2adb5c316e285afa4e19295c2181c0d24" Dec 03 00:19:11 crc kubenswrapper[4953]: I1203 00:19:11.780790 4953 scope.go:117] "RemoveContainer" containerID="00a2ea4a9bb870fce30e7fa845f787f184e7a19e42c95511655ea8c5f84a26d1" Dec 03 00:19:11 crc kubenswrapper[4953]: I1203 00:19:11.795178 4953 scope.go:117] "RemoveContainer" containerID="dc0a5d4cc9d614d8492626a1b114bdc52f510eb5965d842c02c083f08d4849a7" Dec 03 00:19:11 crc kubenswrapper[4953]: I1203 00:19:11.812934 4953 scope.go:117] "RemoveContainer" containerID="af4c7764fca87144d01f3738ea010308744ccf50c54b7c471ccc213112de00d5" Dec 03 00:19:11 crc kubenswrapper[4953]: I1203 00:19:11.861317 4953 scope.go:117] "RemoveContainer" containerID="bcde8ca2a20e2afee0f6cb7164e6068aa82f05c0c978ea3cc4c23618fcb57bac" Dec 03 00:19:11 crc kubenswrapper[4953]: E1203 00:19:11.862173 4953 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bcde8ca2a20e2afee0f6cb7164e6068aa82f05c0c978ea3cc4c23618fcb57bac\": container with ID starting with bcde8ca2a20e2afee0f6cb7164e6068aa82f05c0c978ea3cc4c23618fcb57bac not found: ID does not exist" containerID="bcde8ca2a20e2afee0f6cb7164e6068aa82f05c0c978ea3cc4c23618fcb57bac" Dec 03 00:19:11 crc kubenswrapper[4953]: I1203 00:19:11.862214 4953 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bcde8ca2a20e2afee0f6cb7164e6068aa82f05c0c978ea3cc4c23618fcb57bac"} err="failed to get container status \"bcde8ca2a20e2afee0f6cb7164e6068aa82f05c0c978ea3cc4c23618fcb57bac\": rpc error: code = NotFound desc = could not find container \"bcde8ca2a20e2afee0f6cb7164e6068aa82f05c0c978ea3cc4c23618fcb57bac\": container with ID starting with bcde8ca2a20e2afee0f6cb7164e6068aa82f05c0c978ea3cc4c23618fcb57bac not found: ID does not exist" Dec 03 00:19:11 crc kubenswrapper[4953]: I1203 00:19:11.862244 4953 scope.go:117] "RemoveContainer" containerID="0a4b625eda9b9daff8aa7fe3c5b94e417bcefe08fc6a65b4594f821cfdca7bfd" Dec 03 00:19:11 crc kubenswrapper[4953]: E1203 00:19:11.862806 4953 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0a4b625eda9b9daff8aa7fe3c5b94e417bcefe08fc6a65b4594f821cfdca7bfd\": container with ID starting with 0a4b625eda9b9daff8aa7fe3c5b94e417bcefe08fc6a65b4594f821cfdca7bfd not found: ID does not exist" containerID="0a4b625eda9b9daff8aa7fe3c5b94e417bcefe08fc6a65b4594f821cfdca7bfd" Dec 03 00:19:11 crc kubenswrapper[4953]: I1203 00:19:11.862835 4953 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0a4b625eda9b9daff8aa7fe3c5b94e417bcefe08fc6a65b4594f821cfdca7bfd"} err="failed to get container status \"0a4b625eda9b9daff8aa7fe3c5b94e417bcefe08fc6a65b4594f821cfdca7bfd\": rpc error: code = NotFound desc = could not find container \"0a4b625eda9b9daff8aa7fe3c5b94e417bcefe08fc6a65b4594f821cfdca7bfd\": container with ID starting with 0a4b625eda9b9daff8aa7fe3c5b94e417bcefe08fc6a65b4594f821cfdca7bfd not found: ID does not exist" Dec 03 00:19:11 crc kubenswrapper[4953]: I1203 00:19:11.862857 4953 scope.go:117] "RemoveContainer" containerID="7deadc8e135ff1bc7c5a2f093e489bb5b4b6f166bc874c6acc2a841c5397f11e" Dec 03 00:19:11 crc kubenswrapper[4953]: E1203 00:19:11.863251 4953 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7deadc8e135ff1bc7c5a2f093e489bb5b4b6f166bc874c6acc2a841c5397f11e\": container with ID starting with 7deadc8e135ff1bc7c5a2f093e489bb5b4b6f166bc874c6acc2a841c5397f11e not found: ID does not exist" containerID="7deadc8e135ff1bc7c5a2f093e489bb5b4b6f166bc874c6acc2a841c5397f11e" Dec 03 00:19:11 crc kubenswrapper[4953]: I1203 00:19:11.863284 4953 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7deadc8e135ff1bc7c5a2f093e489bb5b4b6f166bc874c6acc2a841c5397f11e"} err="failed to get container status \"7deadc8e135ff1bc7c5a2f093e489bb5b4b6f166bc874c6acc2a841c5397f11e\": rpc error: code = NotFound desc = could not find container \"7deadc8e135ff1bc7c5a2f093e489bb5b4b6f166bc874c6acc2a841c5397f11e\": container with ID starting with 7deadc8e135ff1bc7c5a2f093e489bb5b4b6f166bc874c6acc2a841c5397f11e not found: ID does not exist" Dec 03 00:19:11 crc kubenswrapper[4953]: I1203 00:19:11.863302 4953 scope.go:117] "RemoveContainer" containerID="a61b951a9cbab2c48c0f2df8351f97d7ff71bf3674f641fe5eac30155232b224" Dec 03 00:19:11 crc kubenswrapper[4953]: E1203 00:19:11.863484 4953 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a61b951a9cbab2c48c0f2df8351f97d7ff71bf3674f641fe5eac30155232b224\": container with ID starting with a61b951a9cbab2c48c0f2df8351f97d7ff71bf3674f641fe5eac30155232b224 not found: ID does not exist" containerID="a61b951a9cbab2c48c0f2df8351f97d7ff71bf3674f641fe5eac30155232b224" Dec 03 00:19:11 crc kubenswrapper[4953]: I1203 00:19:11.863511 4953 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a61b951a9cbab2c48c0f2df8351f97d7ff71bf3674f641fe5eac30155232b224"} err="failed to get container status \"a61b951a9cbab2c48c0f2df8351f97d7ff71bf3674f641fe5eac30155232b224\": rpc error: code = NotFound desc = could not find container \"a61b951a9cbab2c48c0f2df8351f97d7ff71bf3674f641fe5eac30155232b224\": container with ID starting with a61b951a9cbab2c48c0f2df8351f97d7ff71bf3674f641fe5eac30155232b224 not found: ID does not exist" Dec 03 00:19:11 crc kubenswrapper[4953]: I1203 00:19:11.863529 4953 scope.go:117] "RemoveContainer" containerID="fd5f428de229749485267e81d34e31b708728b52474032e88460e54747f21e11" Dec 03 00:19:11 crc kubenswrapper[4953]: E1203 00:19:11.863695 4953 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fd5f428de229749485267e81d34e31b708728b52474032e88460e54747f21e11\": container with ID starting with fd5f428de229749485267e81d34e31b708728b52474032e88460e54747f21e11 not found: ID does not exist" containerID="fd5f428de229749485267e81d34e31b708728b52474032e88460e54747f21e11" Dec 03 00:19:11 crc kubenswrapper[4953]: I1203 00:19:11.863728 4953 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fd5f428de229749485267e81d34e31b708728b52474032e88460e54747f21e11"} err="failed to get container status \"fd5f428de229749485267e81d34e31b708728b52474032e88460e54747f21e11\": rpc error: code = NotFound desc = could not find container \"fd5f428de229749485267e81d34e31b708728b52474032e88460e54747f21e11\": container with ID starting with fd5f428de229749485267e81d34e31b708728b52474032e88460e54747f21e11 not found: ID does not exist" Dec 03 00:19:11 crc kubenswrapper[4953]: I1203 00:19:11.863743 4953 scope.go:117] "RemoveContainer" containerID="e0f6f645a58993f052a6f365bb808f62d99f7fea74886c5cc022b2d86112321a" Dec 03 00:19:11 crc kubenswrapper[4953]: E1203 00:19:11.863914 4953 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e0f6f645a58993f052a6f365bb808f62d99f7fea74886c5cc022b2d86112321a\": container with ID starting with e0f6f645a58993f052a6f365bb808f62d99f7fea74886c5cc022b2d86112321a not found: ID does not exist" containerID="e0f6f645a58993f052a6f365bb808f62d99f7fea74886c5cc022b2d86112321a" Dec 03 00:19:11 crc kubenswrapper[4953]: I1203 00:19:11.863939 4953 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e0f6f645a58993f052a6f365bb808f62d99f7fea74886c5cc022b2d86112321a"} err="failed to get container status \"e0f6f645a58993f052a6f365bb808f62d99f7fea74886c5cc022b2d86112321a\": rpc error: code = NotFound desc = could not find container \"e0f6f645a58993f052a6f365bb808f62d99f7fea74886c5cc022b2d86112321a\": container with ID starting with e0f6f645a58993f052a6f365bb808f62d99f7fea74886c5cc022b2d86112321a not found: ID does not exist" Dec 03 00:19:11 crc kubenswrapper[4953]: I1203 00:19:11.863958 4953 scope.go:117] "RemoveContainer" containerID="d4b85224419a4075acb340e4b11095f2adb5c316e285afa4e19295c2181c0d24" Dec 03 00:19:11 crc kubenswrapper[4953]: E1203 00:19:11.864192 4953 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d4b85224419a4075acb340e4b11095f2adb5c316e285afa4e19295c2181c0d24\": container with ID starting with d4b85224419a4075acb340e4b11095f2adb5c316e285afa4e19295c2181c0d24 not found: ID does not exist" containerID="d4b85224419a4075acb340e4b11095f2adb5c316e285afa4e19295c2181c0d24" Dec 03 00:19:11 crc kubenswrapper[4953]: I1203 00:19:11.864227 4953 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d4b85224419a4075acb340e4b11095f2adb5c316e285afa4e19295c2181c0d24"} err="failed to get container status \"d4b85224419a4075acb340e4b11095f2adb5c316e285afa4e19295c2181c0d24\": rpc error: code = NotFound desc = could not find container \"d4b85224419a4075acb340e4b11095f2adb5c316e285afa4e19295c2181c0d24\": container with ID starting with d4b85224419a4075acb340e4b11095f2adb5c316e285afa4e19295c2181c0d24 not found: ID does not exist" Dec 03 00:19:11 crc kubenswrapper[4953]: I1203 00:19:11.864244 4953 scope.go:117] "RemoveContainer" containerID="00a2ea4a9bb870fce30e7fa845f787f184e7a19e42c95511655ea8c5f84a26d1" Dec 03 00:19:11 crc kubenswrapper[4953]: E1203 00:19:11.864666 4953 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"00a2ea4a9bb870fce30e7fa845f787f184e7a19e42c95511655ea8c5f84a26d1\": container with ID starting with 00a2ea4a9bb870fce30e7fa845f787f184e7a19e42c95511655ea8c5f84a26d1 not found: ID does not exist" containerID="00a2ea4a9bb870fce30e7fa845f787f184e7a19e42c95511655ea8c5f84a26d1" Dec 03 00:19:11 crc kubenswrapper[4953]: I1203 00:19:11.864690 4953 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"00a2ea4a9bb870fce30e7fa845f787f184e7a19e42c95511655ea8c5f84a26d1"} err="failed to get container status \"00a2ea4a9bb870fce30e7fa845f787f184e7a19e42c95511655ea8c5f84a26d1\": rpc error: code = NotFound desc = could not find container \"00a2ea4a9bb870fce30e7fa845f787f184e7a19e42c95511655ea8c5f84a26d1\": container with ID starting with 00a2ea4a9bb870fce30e7fa845f787f184e7a19e42c95511655ea8c5f84a26d1 not found: ID does not exist" Dec 03 00:19:11 crc kubenswrapper[4953]: I1203 00:19:11.864724 4953 scope.go:117] "RemoveContainer" containerID="dc0a5d4cc9d614d8492626a1b114bdc52f510eb5965d842c02c083f08d4849a7" Dec 03 00:19:11 crc kubenswrapper[4953]: E1203 00:19:11.865416 4953 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dc0a5d4cc9d614d8492626a1b114bdc52f510eb5965d842c02c083f08d4849a7\": container with ID starting with dc0a5d4cc9d614d8492626a1b114bdc52f510eb5965d842c02c083f08d4849a7 not found: ID does not exist" containerID="dc0a5d4cc9d614d8492626a1b114bdc52f510eb5965d842c02c083f08d4849a7" Dec 03 00:19:11 crc kubenswrapper[4953]: I1203 00:19:11.865481 4953 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dc0a5d4cc9d614d8492626a1b114bdc52f510eb5965d842c02c083f08d4849a7"} err="failed to get container status \"dc0a5d4cc9d614d8492626a1b114bdc52f510eb5965d842c02c083f08d4849a7\": rpc error: code = NotFound desc = could not find container \"dc0a5d4cc9d614d8492626a1b114bdc52f510eb5965d842c02c083f08d4849a7\": container with ID starting with dc0a5d4cc9d614d8492626a1b114bdc52f510eb5965d842c02c083f08d4849a7 not found: ID does not exist" Dec 03 00:19:11 crc kubenswrapper[4953]: I1203 00:19:11.865522 4953 scope.go:117] "RemoveContainer" containerID="af4c7764fca87144d01f3738ea010308744ccf50c54b7c471ccc213112de00d5" Dec 03 00:19:11 crc kubenswrapper[4953]: E1203 00:19:11.865999 4953 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"af4c7764fca87144d01f3738ea010308744ccf50c54b7c471ccc213112de00d5\": container with ID starting with af4c7764fca87144d01f3738ea010308744ccf50c54b7c471ccc213112de00d5 not found: ID does not exist" containerID="af4c7764fca87144d01f3738ea010308744ccf50c54b7c471ccc213112de00d5" Dec 03 00:19:11 crc kubenswrapper[4953]: I1203 00:19:11.866025 4953 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"af4c7764fca87144d01f3738ea010308744ccf50c54b7c471ccc213112de00d5"} err="failed to get container status \"af4c7764fca87144d01f3738ea010308744ccf50c54b7c471ccc213112de00d5\": rpc error: code = NotFound desc = could not find container \"af4c7764fca87144d01f3738ea010308744ccf50c54b7c471ccc213112de00d5\": container with ID starting with af4c7764fca87144d01f3738ea010308744ccf50c54b7c471ccc213112de00d5 not found: ID does not exist" Dec 03 00:19:11 crc kubenswrapper[4953]: I1203 00:19:11.866042 4953 scope.go:117] "RemoveContainer" containerID="bcde8ca2a20e2afee0f6cb7164e6068aa82f05c0c978ea3cc4c23618fcb57bac" Dec 03 00:19:11 crc kubenswrapper[4953]: I1203 00:19:11.866473 4953 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bcde8ca2a20e2afee0f6cb7164e6068aa82f05c0c978ea3cc4c23618fcb57bac"} err="failed to get container status \"bcde8ca2a20e2afee0f6cb7164e6068aa82f05c0c978ea3cc4c23618fcb57bac\": rpc error: code = NotFound desc = could not find container \"bcde8ca2a20e2afee0f6cb7164e6068aa82f05c0c978ea3cc4c23618fcb57bac\": container with ID starting with bcde8ca2a20e2afee0f6cb7164e6068aa82f05c0c978ea3cc4c23618fcb57bac not found: ID does not exist" Dec 03 00:19:11 crc kubenswrapper[4953]: I1203 00:19:11.866510 4953 scope.go:117] "RemoveContainer" containerID="0a4b625eda9b9daff8aa7fe3c5b94e417bcefe08fc6a65b4594f821cfdca7bfd" Dec 03 00:19:11 crc kubenswrapper[4953]: I1203 00:19:11.866939 4953 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0a4b625eda9b9daff8aa7fe3c5b94e417bcefe08fc6a65b4594f821cfdca7bfd"} err="failed to get container status \"0a4b625eda9b9daff8aa7fe3c5b94e417bcefe08fc6a65b4594f821cfdca7bfd\": rpc error: code = NotFound desc = could not find container \"0a4b625eda9b9daff8aa7fe3c5b94e417bcefe08fc6a65b4594f821cfdca7bfd\": container with ID starting with 0a4b625eda9b9daff8aa7fe3c5b94e417bcefe08fc6a65b4594f821cfdca7bfd not found: ID does not exist" Dec 03 00:19:11 crc kubenswrapper[4953]: I1203 00:19:11.867065 4953 scope.go:117] "RemoveContainer" containerID="7deadc8e135ff1bc7c5a2f093e489bb5b4b6f166bc874c6acc2a841c5397f11e" Dec 03 00:19:11 crc kubenswrapper[4953]: I1203 00:19:11.867507 4953 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7deadc8e135ff1bc7c5a2f093e489bb5b4b6f166bc874c6acc2a841c5397f11e"} err="failed to get container status \"7deadc8e135ff1bc7c5a2f093e489bb5b4b6f166bc874c6acc2a841c5397f11e\": rpc error: code = NotFound desc = could not find container \"7deadc8e135ff1bc7c5a2f093e489bb5b4b6f166bc874c6acc2a841c5397f11e\": container with ID starting with 7deadc8e135ff1bc7c5a2f093e489bb5b4b6f166bc874c6acc2a841c5397f11e not found: ID does not exist" Dec 03 00:19:11 crc kubenswrapper[4953]: I1203 00:19:11.867540 4953 scope.go:117] "RemoveContainer" containerID="a61b951a9cbab2c48c0f2df8351f97d7ff71bf3674f641fe5eac30155232b224" Dec 03 00:19:11 crc kubenswrapper[4953]: I1203 00:19:11.867841 4953 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a61b951a9cbab2c48c0f2df8351f97d7ff71bf3674f641fe5eac30155232b224"} err="failed to get container status \"a61b951a9cbab2c48c0f2df8351f97d7ff71bf3674f641fe5eac30155232b224\": rpc error: code = NotFound desc = could not find container \"a61b951a9cbab2c48c0f2df8351f97d7ff71bf3674f641fe5eac30155232b224\": container with ID starting with a61b951a9cbab2c48c0f2df8351f97d7ff71bf3674f641fe5eac30155232b224 not found: ID does not exist" Dec 03 00:19:11 crc kubenswrapper[4953]: I1203 00:19:11.867867 4953 scope.go:117] "RemoveContainer" containerID="fd5f428de229749485267e81d34e31b708728b52474032e88460e54747f21e11" Dec 03 00:19:11 crc kubenswrapper[4953]: I1203 00:19:11.868141 4953 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fd5f428de229749485267e81d34e31b708728b52474032e88460e54747f21e11"} err="failed to get container status \"fd5f428de229749485267e81d34e31b708728b52474032e88460e54747f21e11\": rpc error: code = NotFound desc = could not find container \"fd5f428de229749485267e81d34e31b708728b52474032e88460e54747f21e11\": container with ID starting with fd5f428de229749485267e81d34e31b708728b52474032e88460e54747f21e11 not found: ID does not exist" Dec 03 00:19:11 crc kubenswrapper[4953]: I1203 00:19:11.868167 4953 scope.go:117] "RemoveContainer" containerID="e0f6f645a58993f052a6f365bb808f62d99f7fea74886c5cc022b2d86112321a" Dec 03 00:19:11 crc kubenswrapper[4953]: I1203 00:19:11.868430 4953 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e0f6f645a58993f052a6f365bb808f62d99f7fea74886c5cc022b2d86112321a"} err="failed to get container status \"e0f6f645a58993f052a6f365bb808f62d99f7fea74886c5cc022b2d86112321a\": rpc error: code = NotFound desc = could not find container \"e0f6f645a58993f052a6f365bb808f62d99f7fea74886c5cc022b2d86112321a\": container with ID starting with e0f6f645a58993f052a6f365bb808f62d99f7fea74886c5cc022b2d86112321a not found: ID does not exist" Dec 03 00:19:11 crc kubenswrapper[4953]: I1203 00:19:11.868451 4953 scope.go:117] "RemoveContainer" containerID="d4b85224419a4075acb340e4b11095f2adb5c316e285afa4e19295c2181c0d24" Dec 03 00:19:11 crc kubenswrapper[4953]: I1203 00:19:11.868716 4953 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d4b85224419a4075acb340e4b11095f2adb5c316e285afa4e19295c2181c0d24"} err="failed to get container status \"d4b85224419a4075acb340e4b11095f2adb5c316e285afa4e19295c2181c0d24\": rpc error: code = NotFound desc = could not find container \"d4b85224419a4075acb340e4b11095f2adb5c316e285afa4e19295c2181c0d24\": container with ID starting with d4b85224419a4075acb340e4b11095f2adb5c316e285afa4e19295c2181c0d24 not found: ID does not exist" Dec 03 00:19:11 crc kubenswrapper[4953]: I1203 00:19:11.868739 4953 scope.go:117] "RemoveContainer" containerID="00a2ea4a9bb870fce30e7fa845f787f184e7a19e42c95511655ea8c5f84a26d1" Dec 03 00:19:11 crc kubenswrapper[4953]: I1203 00:19:11.868965 4953 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"00a2ea4a9bb870fce30e7fa845f787f184e7a19e42c95511655ea8c5f84a26d1"} err="failed to get container status \"00a2ea4a9bb870fce30e7fa845f787f184e7a19e42c95511655ea8c5f84a26d1\": rpc error: code = NotFound desc = could not find container \"00a2ea4a9bb870fce30e7fa845f787f184e7a19e42c95511655ea8c5f84a26d1\": container with ID starting with 00a2ea4a9bb870fce30e7fa845f787f184e7a19e42c95511655ea8c5f84a26d1 not found: ID does not exist" Dec 03 00:19:11 crc kubenswrapper[4953]: I1203 00:19:11.868999 4953 scope.go:117] "RemoveContainer" containerID="dc0a5d4cc9d614d8492626a1b114bdc52f510eb5965d842c02c083f08d4849a7" Dec 03 00:19:11 crc kubenswrapper[4953]: I1203 00:19:11.869403 4953 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dc0a5d4cc9d614d8492626a1b114bdc52f510eb5965d842c02c083f08d4849a7"} err="failed to get container status \"dc0a5d4cc9d614d8492626a1b114bdc52f510eb5965d842c02c083f08d4849a7\": rpc error: code = NotFound desc = could not find container \"dc0a5d4cc9d614d8492626a1b114bdc52f510eb5965d842c02c083f08d4849a7\": container with ID starting with dc0a5d4cc9d614d8492626a1b114bdc52f510eb5965d842c02c083f08d4849a7 not found: ID does not exist" Dec 03 00:19:11 crc kubenswrapper[4953]: I1203 00:19:11.869436 4953 scope.go:117] "RemoveContainer" containerID="af4c7764fca87144d01f3738ea010308744ccf50c54b7c471ccc213112de00d5" Dec 03 00:19:11 crc kubenswrapper[4953]: I1203 00:19:11.869821 4953 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"af4c7764fca87144d01f3738ea010308744ccf50c54b7c471ccc213112de00d5"} err="failed to get container status \"af4c7764fca87144d01f3738ea010308744ccf50c54b7c471ccc213112de00d5\": rpc error: code = NotFound desc = could not find container \"af4c7764fca87144d01f3738ea010308744ccf50c54b7c471ccc213112de00d5\": container with ID starting with af4c7764fca87144d01f3738ea010308744ccf50c54b7c471ccc213112de00d5 not found: ID does not exist" Dec 03 00:19:11 crc kubenswrapper[4953]: I1203 00:19:11.869842 4953 scope.go:117] "RemoveContainer" containerID="bcde8ca2a20e2afee0f6cb7164e6068aa82f05c0c978ea3cc4c23618fcb57bac" Dec 03 00:19:11 crc kubenswrapper[4953]: I1203 00:19:11.870183 4953 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bcde8ca2a20e2afee0f6cb7164e6068aa82f05c0c978ea3cc4c23618fcb57bac"} err="failed to get container status \"bcde8ca2a20e2afee0f6cb7164e6068aa82f05c0c978ea3cc4c23618fcb57bac\": rpc error: code = NotFound desc = could not find container \"bcde8ca2a20e2afee0f6cb7164e6068aa82f05c0c978ea3cc4c23618fcb57bac\": container with ID starting with bcde8ca2a20e2afee0f6cb7164e6068aa82f05c0c978ea3cc4c23618fcb57bac not found: ID does not exist" Dec 03 00:19:11 crc kubenswrapper[4953]: I1203 00:19:11.870208 4953 scope.go:117] "RemoveContainer" containerID="0a4b625eda9b9daff8aa7fe3c5b94e417bcefe08fc6a65b4594f821cfdca7bfd" Dec 03 00:19:11 crc kubenswrapper[4953]: I1203 00:19:11.870529 4953 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0a4b625eda9b9daff8aa7fe3c5b94e417bcefe08fc6a65b4594f821cfdca7bfd"} err="failed to get container status \"0a4b625eda9b9daff8aa7fe3c5b94e417bcefe08fc6a65b4594f821cfdca7bfd\": rpc error: code = NotFound desc = could not find container \"0a4b625eda9b9daff8aa7fe3c5b94e417bcefe08fc6a65b4594f821cfdca7bfd\": container with ID starting with 0a4b625eda9b9daff8aa7fe3c5b94e417bcefe08fc6a65b4594f821cfdca7bfd not found: ID does not exist" Dec 03 00:19:11 crc kubenswrapper[4953]: I1203 00:19:11.870562 4953 scope.go:117] "RemoveContainer" containerID="7deadc8e135ff1bc7c5a2f093e489bb5b4b6f166bc874c6acc2a841c5397f11e" Dec 03 00:19:11 crc kubenswrapper[4953]: I1203 00:19:11.871045 4953 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7deadc8e135ff1bc7c5a2f093e489bb5b4b6f166bc874c6acc2a841c5397f11e"} err="failed to get container status \"7deadc8e135ff1bc7c5a2f093e489bb5b4b6f166bc874c6acc2a841c5397f11e\": rpc error: code = NotFound desc = could not find container \"7deadc8e135ff1bc7c5a2f093e489bb5b4b6f166bc874c6acc2a841c5397f11e\": container with ID starting with 7deadc8e135ff1bc7c5a2f093e489bb5b4b6f166bc874c6acc2a841c5397f11e not found: ID does not exist" Dec 03 00:19:11 crc kubenswrapper[4953]: I1203 00:19:11.871073 4953 scope.go:117] "RemoveContainer" containerID="a61b951a9cbab2c48c0f2df8351f97d7ff71bf3674f641fe5eac30155232b224" Dec 03 00:19:11 crc kubenswrapper[4953]: I1203 00:19:11.871391 4953 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a61b951a9cbab2c48c0f2df8351f97d7ff71bf3674f641fe5eac30155232b224"} err="failed to get container status \"a61b951a9cbab2c48c0f2df8351f97d7ff71bf3674f641fe5eac30155232b224\": rpc error: code = NotFound desc = could not find container \"a61b951a9cbab2c48c0f2df8351f97d7ff71bf3674f641fe5eac30155232b224\": container with ID starting with a61b951a9cbab2c48c0f2df8351f97d7ff71bf3674f641fe5eac30155232b224 not found: ID does not exist" Dec 03 00:19:11 crc kubenswrapper[4953]: I1203 00:19:11.871408 4953 scope.go:117] "RemoveContainer" containerID="fd5f428de229749485267e81d34e31b708728b52474032e88460e54747f21e11" Dec 03 00:19:11 crc kubenswrapper[4953]: I1203 00:19:11.871675 4953 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fd5f428de229749485267e81d34e31b708728b52474032e88460e54747f21e11"} err="failed to get container status \"fd5f428de229749485267e81d34e31b708728b52474032e88460e54747f21e11\": rpc error: code = NotFound desc = could not find container \"fd5f428de229749485267e81d34e31b708728b52474032e88460e54747f21e11\": container with ID starting with fd5f428de229749485267e81d34e31b708728b52474032e88460e54747f21e11 not found: ID does not exist" Dec 03 00:19:11 crc kubenswrapper[4953]: I1203 00:19:11.871691 4953 scope.go:117] "RemoveContainer" containerID="e0f6f645a58993f052a6f365bb808f62d99f7fea74886c5cc022b2d86112321a" Dec 03 00:19:11 crc kubenswrapper[4953]: I1203 00:19:11.873266 4953 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e0f6f645a58993f052a6f365bb808f62d99f7fea74886c5cc022b2d86112321a"} err="failed to get container status \"e0f6f645a58993f052a6f365bb808f62d99f7fea74886c5cc022b2d86112321a\": rpc error: code = NotFound desc = could not find container \"e0f6f645a58993f052a6f365bb808f62d99f7fea74886c5cc022b2d86112321a\": container with ID starting with e0f6f645a58993f052a6f365bb808f62d99f7fea74886c5cc022b2d86112321a not found: ID does not exist" Dec 03 00:19:11 crc kubenswrapper[4953]: I1203 00:19:11.873314 4953 scope.go:117] "RemoveContainer" containerID="d4b85224419a4075acb340e4b11095f2adb5c316e285afa4e19295c2181c0d24" Dec 03 00:19:11 crc kubenswrapper[4953]: I1203 00:19:11.873660 4953 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d4b85224419a4075acb340e4b11095f2adb5c316e285afa4e19295c2181c0d24"} err="failed to get container status \"d4b85224419a4075acb340e4b11095f2adb5c316e285afa4e19295c2181c0d24\": rpc error: code = NotFound desc = could not find container \"d4b85224419a4075acb340e4b11095f2adb5c316e285afa4e19295c2181c0d24\": container with ID starting with d4b85224419a4075acb340e4b11095f2adb5c316e285afa4e19295c2181c0d24 not found: ID does not exist" Dec 03 00:19:11 crc kubenswrapper[4953]: I1203 00:19:11.873688 4953 scope.go:117] "RemoveContainer" containerID="00a2ea4a9bb870fce30e7fa845f787f184e7a19e42c95511655ea8c5f84a26d1" Dec 03 00:19:11 crc kubenswrapper[4953]: I1203 00:19:11.874028 4953 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"00a2ea4a9bb870fce30e7fa845f787f184e7a19e42c95511655ea8c5f84a26d1"} err="failed to get container status \"00a2ea4a9bb870fce30e7fa845f787f184e7a19e42c95511655ea8c5f84a26d1\": rpc error: code = NotFound desc = could not find container \"00a2ea4a9bb870fce30e7fa845f787f184e7a19e42c95511655ea8c5f84a26d1\": container with ID starting with 00a2ea4a9bb870fce30e7fa845f787f184e7a19e42c95511655ea8c5f84a26d1 not found: ID does not exist" Dec 03 00:19:11 crc kubenswrapper[4953]: I1203 00:19:11.874052 4953 scope.go:117] "RemoveContainer" containerID="dc0a5d4cc9d614d8492626a1b114bdc52f510eb5965d842c02c083f08d4849a7" Dec 03 00:19:11 crc kubenswrapper[4953]: I1203 00:19:11.874374 4953 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dc0a5d4cc9d614d8492626a1b114bdc52f510eb5965d842c02c083f08d4849a7"} err="failed to get container status \"dc0a5d4cc9d614d8492626a1b114bdc52f510eb5965d842c02c083f08d4849a7\": rpc error: code = NotFound desc = could not find container \"dc0a5d4cc9d614d8492626a1b114bdc52f510eb5965d842c02c083f08d4849a7\": container with ID starting with dc0a5d4cc9d614d8492626a1b114bdc52f510eb5965d842c02c083f08d4849a7 not found: ID does not exist" Dec 03 00:19:11 crc kubenswrapper[4953]: I1203 00:19:11.874394 4953 scope.go:117] "RemoveContainer" containerID="af4c7764fca87144d01f3738ea010308744ccf50c54b7c471ccc213112de00d5" Dec 03 00:19:11 crc kubenswrapper[4953]: I1203 00:19:11.874744 4953 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"af4c7764fca87144d01f3738ea010308744ccf50c54b7c471ccc213112de00d5"} err="failed to get container status \"af4c7764fca87144d01f3738ea010308744ccf50c54b7c471ccc213112de00d5\": rpc error: code = NotFound desc = could not find container \"af4c7764fca87144d01f3738ea010308744ccf50c54b7c471ccc213112de00d5\": container with ID starting with af4c7764fca87144d01f3738ea010308744ccf50c54b7c471ccc213112de00d5 not found: ID does not exist" Dec 03 00:19:11 crc kubenswrapper[4953]: I1203 00:19:11.874771 4953 scope.go:117] "RemoveContainer" containerID="bcde8ca2a20e2afee0f6cb7164e6068aa82f05c0c978ea3cc4c23618fcb57bac" Dec 03 00:19:11 crc kubenswrapper[4953]: I1203 00:19:11.875165 4953 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bcde8ca2a20e2afee0f6cb7164e6068aa82f05c0c978ea3cc4c23618fcb57bac"} err="failed to get container status \"bcde8ca2a20e2afee0f6cb7164e6068aa82f05c0c978ea3cc4c23618fcb57bac\": rpc error: code = NotFound desc = could not find container \"bcde8ca2a20e2afee0f6cb7164e6068aa82f05c0c978ea3cc4c23618fcb57bac\": container with ID starting with bcde8ca2a20e2afee0f6cb7164e6068aa82f05c0c978ea3cc4c23618fcb57bac not found: ID does not exist" Dec 03 00:19:11 crc kubenswrapper[4953]: I1203 00:19:11.875194 4953 scope.go:117] "RemoveContainer" containerID="0a4b625eda9b9daff8aa7fe3c5b94e417bcefe08fc6a65b4594f821cfdca7bfd" Dec 03 00:19:11 crc kubenswrapper[4953]: I1203 00:19:11.875545 4953 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0a4b625eda9b9daff8aa7fe3c5b94e417bcefe08fc6a65b4594f821cfdca7bfd"} err="failed to get container status \"0a4b625eda9b9daff8aa7fe3c5b94e417bcefe08fc6a65b4594f821cfdca7bfd\": rpc error: code = NotFound desc = could not find container \"0a4b625eda9b9daff8aa7fe3c5b94e417bcefe08fc6a65b4594f821cfdca7bfd\": container with ID starting with 0a4b625eda9b9daff8aa7fe3c5b94e417bcefe08fc6a65b4594f821cfdca7bfd not found: ID does not exist" Dec 03 00:19:11 crc kubenswrapper[4953]: I1203 00:19:11.875574 4953 scope.go:117] "RemoveContainer" containerID="7deadc8e135ff1bc7c5a2f093e489bb5b4b6f166bc874c6acc2a841c5397f11e" Dec 03 00:19:11 crc kubenswrapper[4953]: I1203 00:19:11.875900 4953 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7deadc8e135ff1bc7c5a2f093e489bb5b4b6f166bc874c6acc2a841c5397f11e"} err="failed to get container status \"7deadc8e135ff1bc7c5a2f093e489bb5b4b6f166bc874c6acc2a841c5397f11e\": rpc error: code = NotFound desc = could not find container \"7deadc8e135ff1bc7c5a2f093e489bb5b4b6f166bc874c6acc2a841c5397f11e\": container with ID starting with 7deadc8e135ff1bc7c5a2f093e489bb5b4b6f166bc874c6acc2a841c5397f11e not found: ID does not exist" Dec 03 00:19:11 crc kubenswrapper[4953]: I1203 00:19:11.875924 4953 scope.go:117] "RemoveContainer" containerID="a61b951a9cbab2c48c0f2df8351f97d7ff71bf3674f641fe5eac30155232b224" Dec 03 00:19:11 crc kubenswrapper[4953]: I1203 00:19:11.876465 4953 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a61b951a9cbab2c48c0f2df8351f97d7ff71bf3674f641fe5eac30155232b224"} err="failed to get container status \"a61b951a9cbab2c48c0f2df8351f97d7ff71bf3674f641fe5eac30155232b224\": rpc error: code = NotFound desc = could not find container \"a61b951a9cbab2c48c0f2df8351f97d7ff71bf3674f641fe5eac30155232b224\": container with ID starting with a61b951a9cbab2c48c0f2df8351f97d7ff71bf3674f641fe5eac30155232b224 not found: ID does not exist" Dec 03 00:19:11 crc kubenswrapper[4953]: I1203 00:19:11.876522 4953 scope.go:117] "RemoveContainer" containerID="fd5f428de229749485267e81d34e31b708728b52474032e88460e54747f21e11" Dec 03 00:19:11 crc kubenswrapper[4953]: I1203 00:19:11.876856 4953 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fd5f428de229749485267e81d34e31b708728b52474032e88460e54747f21e11"} err="failed to get container status \"fd5f428de229749485267e81d34e31b708728b52474032e88460e54747f21e11\": rpc error: code = NotFound desc = could not find container \"fd5f428de229749485267e81d34e31b708728b52474032e88460e54747f21e11\": container with ID starting with fd5f428de229749485267e81d34e31b708728b52474032e88460e54747f21e11 not found: ID does not exist" Dec 03 00:19:11 crc kubenswrapper[4953]: I1203 00:19:11.876875 4953 scope.go:117] "RemoveContainer" containerID="e0f6f645a58993f052a6f365bb808f62d99f7fea74886c5cc022b2d86112321a" Dec 03 00:19:11 crc kubenswrapper[4953]: I1203 00:19:11.877250 4953 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e0f6f645a58993f052a6f365bb808f62d99f7fea74886c5cc022b2d86112321a"} err="failed to get container status \"e0f6f645a58993f052a6f365bb808f62d99f7fea74886c5cc022b2d86112321a\": rpc error: code = NotFound desc = could not find container \"e0f6f645a58993f052a6f365bb808f62d99f7fea74886c5cc022b2d86112321a\": container with ID starting with e0f6f645a58993f052a6f365bb808f62d99f7fea74886c5cc022b2d86112321a not found: ID does not exist" Dec 03 00:19:11 crc kubenswrapper[4953]: I1203 00:19:11.877281 4953 scope.go:117] "RemoveContainer" containerID="d4b85224419a4075acb340e4b11095f2adb5c316e285afa4e19295c2181c0d24" Dec 03 00:19:11 crc kubenswrapper[4953]: I1203 00:19:11.877581 4953 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d4b85224419a4075acb340e4b11095f2adb5c316e285afa4e19295c2181c0d24"} err="failed to get container status \"d4b85224419a4075acb340e4b11095f2adb5c316e285afa4e19295c2181c0d24\": rpc error: code = NotFound desc = could not find container \"d4b85224419a4075acb340e4b11095f2adb5c316e285afa4e19295c2181c0d24\": container with ID starting with d4b85224419a4075acb340e4b11095f2adb5c316e285afa4e19295c2181c0d24 not found: ID does not exist" Dec 03 00:19:11 crc kubenswrapper[4953]: I1203 00:19:11.877605 4953 scope.go:117] "RemoveContainer" containerID="00a2ea4a9bb870fce30e7fa845f787f184e7a19e42c95511655ea8c5f84a26d1" Dec 03 00:19:11 crc kubenswrapper[4953]: I1203 00:19:11.877930 4953 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"00a2ea4a9bb870fce30e7fa845f787f184e7a19e42c95511655ea8c5f84a26d1"} err="failed to get container status \"00a2ea4a9bb870fce30e7fa845f787f184e7a19e42c95511655ea8c5f84a26d1\": rpc error: code = NotFound desc = could not find container \"00a2ea4a9bb870fce30e7fa845f787f184e7a19e42c95511655ea8c5f84a26d1\": container with ID starting with 00a2ea4a9bb870fce30e7fa845f787f184e7a19e42c95511655ea8c5f84a26d1 not found: ID does not exist" Dec 03 00:19:11 crc kubenswrapper[4953]: I1203 00:19:11.877966 4953 scope.go:117] "RemoveContainer" containerID="dc0a5d4cc9d614d8492626a1b114bdc52f510eb5965d842c02c083f08d4849a7" Dec 03 00:19:11 crc kubenswrapper[4953]: I1203 00:19:11.878470 4953 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dc0a5d4cc9d614d8492626a1b114bdc52f510eb5965d842c02c083f08d4849a7"} err="failed to get container status \"dc0a5d4cc9d614d8492626a1b114bdc52f510eb5965d842c02c083f08d4849a7\": rpc error: code = NotFound desc = could not find container \"dc0a5d4cc9d614d8492626a1b114bdc52f510eb5965d842c02c083f08d4849a7\": container with ID starting with dc0a5d4cc9d614d8492626a1b114bdc52f510eb5965d842c02c083f08d4849a7 not found: ID does not exist" Dec 03 00:19:11 crc kubenswrapper[4953]: I1203 00:19:11.878499 4953 scope.go:117] "RemoveContainer" containerID="af4c7764fca87144d01f3738ea010308744ccf50c54b7c471ccc213112de00d5" Dec 03 00:19:11 crc kubenswrapper[4953]: I1203 00:19:11.879015 4953 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"af4c7764fca87144d01f3738ea010308744ccf50c54b7c471ccc213112de00d5"} err="failed to get container status \"af4c7764fca87144d01f3738ea010308744ccf50c54b7c471ccc213112de00d5\": rpc error: code = NotFound desc = could not find container \"af4c7764fca87144d01f3738ea010308744ccf50c54b7c471ccc213112de00d5\": container with ID starting with af4c7764fca87144d01f3738ea010308744ccf50c54b7c471ccc213112de00d5 not found: ID does not exist" Dec 03 00:19:11 crc kubenswrapper[4953]: I1203 00:19:11.879039 4953 scope.go:117] "RemoveContainer" containerID="bcde8ca2a20e2afee0f6cb7164e6068aa82f05c0c978ea3cc4c23618fcb57bac" Dec 03 00:19:11 crc kubenswrapper[4953]: I1203 00:19:11.879388 4953 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bcde8ca2a20e2afee0f6cb7164e6068aa82f05c0c978ea3cc4c23618fcb57bac"} err="failed to get container status \"bcde8ca2a20e2afee0f6cb7164e6068aa82f05c0c978ea3cc4c23618fcb57bac\": rpc error: code = NotFound desc = could not find container \"bcde8ca2a20e2afee0f6cb7164e6068aa82f05c0c978ea3cc4c23618fcb57bac\": container with ID starting with bcde8ca2a20e2afee0f6cb7164e6068aa82f05c0c978ea3cc4c23618fcb57bac not found: ID does not exist" Dec 03 00:19:12 crc kubenswrapper[4953]: I1203 00:19:12.598211 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fnl4b" event={"ID":"614357c4-4ad5-4c35-9876-7e4e8ded1c9c","Type":"ContainerStarted","Data":"a5e895b1b82dedf15fd6eb67d76195c3527257d9265d81d1bbf77eb5bc99df47"} Dec 03 00:19:12 crc kubenswrapper[4953]: I1203 00:19:12.598274 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fnl4b" event={"ID":"614357c4-4ad5-4c35-9876-7e4e8ded1c9c","Type":"ContainerStarted","Data":"512dcad30bcd611137e27f6dea6b2ef4082b5c6d0b22ec6db7083b37342b2ec1"} Dec 03 00:19:12 crc kubenswrapper[4953]: I1203 00:19:12.598290 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fnl4b" event={"ID":"614357c4-4ad5-4c35-9876-7e4e8ded1c9c","Type":"ContainerStarted","Data":"1a4b372e11f5c6d232f81cc9b97f5fd079885c931756284e5c4ca90a8205e4b9"} Dec 03 00:19:12 crc kubenswrapper[4953]: I1203 00:19:12.598303 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fnl4b" event={"ID":"614357c4-4ad5-4c35-9876-7e4e8ded1c9c","Type":"ContainerStarted","Data":"8313e81fc37921f6dea3e0b4c5d3835fadccc8c9dc18cfa5d5d936087190c004"} Dec 03 00:19:12 crc kubenswrapper[4953]: I1203 00:19:12.598313 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fnl4b" event={"ID":"614357c4-4ad5-4c35-9876-7e4e8ded1c9c","Type":"ContainerStarted","Data":"e74681ea7300c0412ae4e14685188d1bdf4c5f259c1f3500923021715f594d1d"} Dec 03 00:19:12 crc kubenswrapper[4953]: I1203 00:19:12.598323 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fnl4b" event={"ID":"614357c4-4ad5-4c35-9876-7e4e8ded1c9c","Type":"ContainerStarted","Data":"6c8279d5c214a8c7e51ca49accde9d3f031e9a28e85e86400fa6c7eca389b9df"} Dec 03 00:19:13 crc kubenswrapper[4953]: I1203 00:19:13.359907 4953 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8816ebee-c735-4b8d-b1ec-65349707b347" path="/var/lib/kubelet/pods/8816ebee-c735-4b8d-b1ec-65349707b347/volumes" Dec 03 00:19:15 crc kubenswrapper[4953]: I1203 00:19:15.623517 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fnl4b" event={"ID":"614357c4-4ad5-4c35-9876-7e4e8ded1c9c","Type":"ContainerStarted","Data":"68e52aa89fc66de0bf2b0be118f6e3fe246f917788fe515083703aef82c0d2ee"} Dec 03 00:19:17 crc kubenswrapper[4953]: I1203 00:19:17.645651 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fnl4b" event={"ID":"614357c4-4ad5-4c35-9876-7e4e8ded1c9c","Type":"ContainerStarted","Data":"f8bb4bb840d3d18370960fcc19546af4d2fd407bfa0ed4eeecf8597ee5225532"} Dec 03 00:19:17 crc kubenswrapper[4953]: I1203 00:19:17.646124 4953 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-fnl4b" Dec 03 00:19:17 crc kubenswrapper[4953]: I1203 00:19:17.646225 4953 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-fnl4b" Dec 03 00:19:17 crc kubenswrapper[4953]: I1203 00:19:17.680200 4953 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-fnl4b" podStartSLOduration=7.680173194 podStartE2EDuration="7.680173194s" podCreationTimestamp="2025-12-03 00:19:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 00:19:17.671502289 +0000 UTC m=+746.655783092" watchObservedRunningTime="2025-12-03 00:19:17.680173194 +0000 UTC m=+746.664453977" Dec 03 00:19:17 crc kubenswrapper[4953]: I1203 00:19:17.691785 4953 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-fnl4b" Dec 03 00:19:18 crc kubenswrapper[4953]: I1203 00:19:18.654153 4953 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-fnl4b" Dec 03 00:19:18 crc kubenswrapper[4953]: I1203 00:19:18.690847 4953 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-fnl4b" Dec 03 00:19:34 crc kubenswrapper[4953]: I1203 00:19:34.517710 4953 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Dec 03 00:19:41 crc kubenswrapper[4953]: I1203 00:19:41.264398 4953 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-fnl4b" Dec 03 00:19:48 crc kubenswrapper[4953]: I1203 00:19:48.944770 4953 patch_prober.go:28] interesting pod/machine-config-daemon-6dfbn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 00:19:48 crc kubenswrapper[4953]: I1203 00:19:48.945745 4953 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-6dfbn" podUID="53ffb1e9-9a2e-4869-80d7-776f4aa7ef8b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 00:20:18 crc kubenswrapper[4953]: I1203 00:20:18.944721 4953 patch_prober.go:28] interesting pod/machine-config-daemon-6dfbn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 00:20:18 crc kubenswrapper[4953]: I1203 00:20:18.945788 4953 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-6dfbn" podUID="53ffb1e9-9a2e-4869-80d7-776f4aa7ef8b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 00:20:23 crc kubenswrapper[4953]: I1203 00:20:23.370423 4953 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-dpsq6"] Dec 03 00:20:23 crc kubenswrapper[4953]: I1203 00:20:23.371191 4953 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-dpsq6" podUID="c2b69c71-8e67-461d-9ddf-cfb439ebb6e1" containerName="registry-server" containerID="cri-o://034c2b2807e39fd7848b03f21692b9a4094c04152d8464a2ca8f5d82217eb39a" gracePeriod=30 Dec 03 00:20:23 crc kubenswrapper[4953]: I1203 00:20:23.799072 4953 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-dpsq6" Dec 03 00:20:23 crc kubenswrapper[4953]: I1203 00:20:23.934783 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c2b69c71-8e67-461d-9ddf-cfb439ebb6e1-catalog-content\") pod \"c2b69c71-8e67-461d-9ddf-cfb439ebb6e1\" (UID: \"c2b69c71-8e67-461d-9ddf-cfb439ebb6e1\") " Dec 03 00:20:23 crc kubenswrapper[4953]: I1203 00:20:23.934879 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c2b69c71-8e67-461d-9ddf-cfb439ebb6e1-utilities\") pod \"c2b69c71-8e67-461d-9ddf-cfb439ebb6e1\" (UID: \"c2b69c71-8e67-461d-9ddf-cfb439ebb6e1\") " Dec 03 00:20:23 crc kubenswrapper[4953]: I1203 00:20:23.934949 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gb6nx\" (UniqueName: \"kubernetes.io/projected/c2b69c71-8e67-461d-9ddf-cfb439ebb6e1-kube-api-access-gb6nx\") pod \"c2b69c71-8e67-461d-9ddf-cfb439ebb6e1\" (UID: \"c2b69c71-8e67-461d-9ddf-cfb439ebb6e1\") " Dec 03 00:20:23 crc kubenswrapper[4953]: I1203 00:20:23.936280 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c2b69c71-8e67-461d-9ddf-cfb439ebb6e1-utilities" (OuterVolumeSpecName: "utilities") pod "c2b69c71-8e67-461d-9ddf-cfb439ebb6e1" (UID: "c2b69c71-8e67-461d-9ddf-cfb439ebb6e1"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 00:20:23 crc kubenswrapper[4953]: I1203 00:20:23.942465 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c2b69c71-8e67-461d-9ddf-cfb439ebb6e1-kube-api-access-gb6nx" (OuterVolumeSpecName: "kube-api-access-gb6nx") pod "c2b69c71-8e67-461d-9ddf-cfb439ebb6e1" (UID: "c2b69c71-8e67-461d-9ddf-cfb439ebb6e1"). InnerVolumeSpecName "kube-api-access-gb6nx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:20:23 crc kubenswrapper[4953]: I1203 00:20:23.954293 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c2b69c71-8e67-461d-9ddf-cfb439ebb6e1-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c2b69c71-8e67-461d-9ddf-cfb439ebb6e1" (UID: "c2b69c71-8e67-461d-9ddf-cfb439ebb6e1"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 00:20:24 crc kubenswrapper[4953]: I1203 00:20:24.037056 4953 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c2b69c71-8e67-461d-9ddf-cfb439ebb6e1-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 00:20:24 crc kubenswrapper[4953]: I1203 00:20:24.037114 4953 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c2b69c71-8e67-461d-9ddf-cfb439ebb6e1-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 00:20:24 crc kubenswrapper[4953]: I1203 00:20:24.037140 4953 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gb6nx\" (UniqueName: \"kubernetes.io/projected/c2b69c71-8e67-461d-9ddf-cfb439ebb6e1-kube-api-access-gb6nx\") on node \"crc\" DevicePath \"\"" Dec 03 00:20:24 crc kubenswrapper[4953]: I1203 00:20:24.067157 4953 generic.go:334] "Generic (PLEG): container finished" podID="c2b69c71-8e67-461d-9ddf-cfb439ebb6e1" containerID="034c2b2807e39fd7848b03f21692b9a4094c04152d8464a2ca8f5d82217eb39a" exitCode=0 Dec 03 00:20:24 crc kubenswrapper[4953]: I1203 00:20:24.067223 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dpsq6" event={"ID":"c2b69c71-8e67-461d-9ddf-cfb439ebb6e1","Type":"ContainerDied","Data":"034c2b2807e39fd7848b03f21692b9a4094c04152d8464a2ca8f5d82217eb39a"} Dec 03 00:20:24 crc kubenswrapper[4953]: I1203 00:20:24.067261 4953 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-dpsq6" Dec 03 00:20:24 crc kubenswrapper[4953]: I1203 00:20:24.067269 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dpsq6" event={"ID":"c2b69c71-8e67-461d-9ddf-cfb439ebb6e1","Type":"ContainerDied","Data":"199c1982390394e5cde47a5fbe1309d30350e3d7d92ac8ca558ca49c7b125c0b"} Dec 03 00:20:24 crc kubenswrapper[4953]: I1203 00:20:24.067306 4953 scope.go:117] "RemoveContainer" containerID="034c2b2807e39fd7848b03f21692b9a4094c04152d8464a2ca8f5d82217eb39a" Dec 03 00:20:24 crc kubenswrapper[4953]: I1203 00:20:24.087944 4953 scope.go:117] "RemoveContainer" containerID="bc8121247f710ab2d28dcd715750c2c07d278ab469c3c0361138ab85f78b76ca" Dec 03 00:20:24 crc kubenswrapper[4953]: I1203 00:20:24.106869 4953 scope.go:117] "RemoveContainer" containerID="b915762a27acbf96f1aad344c6de71c18cb55d5309f14ee44012759acda59245" Dec 03 00:20:24 crc kubenswrapper[4953]: I1203 00:20:24.109525 4953 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-dpsq6"] Dec 03 00:20:24 crc kubenswrapper[4953]: I1203 00:20:24.112866 4953 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-dpsq6"] Dec 03 00:20:24 crc kubenswrapper[4953]: I1203 00:20:24.124268 4953 scope.go:117] "RemoveContainer" containerID="034c2b2807e39fd7848b03f21692b9a4094c04152d8464a2ca8f5d82217eb39a" Dec 03 00:20:24 crc kubenswrapper[4953]: E1203 00:20:24.124759 4953 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"034c2b2807e39fd7848b03f21692b9a4094c04152d8464a2ca8f5d82217eb39a\": container with ID starting with 034c2b2807e39fd7848b03f21692b9a4094c04152d8464a2ca8f5d82217eb39a not found: ID does not exist" containerID="034c2b2807e39fd7848b03f21692b9a4094c04152d8464a2ca8f5d82217eb39a" Dec 03 00:20:24 crc kubenswrapper[4953]: I1203 00:20:24.124792 4953 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"034c2b2807e39fd7848b03f21692b9a4094c04152d8464a2ca8f5d82217eb39a"} err="failed to get container status \"034c2b2807e39fd7848b03f21692b9a4094c04152d8464a2ca8f5d82217eb39a\": rpc error: code = NotFound desc = could not find container \"034c2b2807e39fd7848b03f21692b9a4094c04152d8464a2ca8f5d82217eb39a\": container with ID starting with 034c2b2807e39fd7848b03f21692b9a4094c04152d8464a2ca8f5d82217eb39a not found: ID does not exist" Dec 03 00:20:24 crc kubenswrapper[4953]: I1203 00:20:24.124814 4953 scope.go:117] "RemoveContainer" containerID="bc8121247f710ab2d28dcd715750c2c07d278ab469c3c0361138ab85f78b76ca" Dec 03 00:20:24 crc kubenswrapper[4953]: E1203 00:20:24.125159 4953 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bc8121247f710ab2d28dcd715750c2c07d278ab469c3c0361138ab85f78b76ca\": container with ID starting with bc8121247f710ab2d28dcd715750c2c07d278ab469c3c0361138ab85f78b76ca not found: ID does not exist" containerID="bc8121247f710ab2d28dcd715750c2c07d278ab469c3c0361138ab85f78b76ca" Dec 03 00:20:24 crc kubenswrapper[4953]: I1203 00:20:24.125183 4953 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bc8121247f710ab2d28dcd715750c2c07d278ab469c3c0361138ab85f78b76ca"} err="failed to get container status \"bc8121247f710ab2d28dcd715750c2c07d278ab469c3c0361138ab85f78b76ca\": rpc error: code = NotFound desc = could not find container \"bc8121247f710ab2d28dcd715750c2c07d278ab469c3c0361138ab85f78b76ca\": container with ID starting with bc8121247f710ab2d28dcd715750c2c07d278ab469c3c0361138ab85f78b76ca not found: ID does not exist" Dec 03 00:20:24 crc kubenswrapper[4953]: I1203 00:20:24.125198 4953 scope.go:117] "RemoveContainer" containerID="b915762a27acbf96f1aad344c6de71c18cb55d5309f14ee44012759acda59245" Dec 03 00:20:24 crc kubenswrapper[4953]: E1203 00:20:24.125712 4953 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b915762a27acbf96f1aad344c6de71c18cb55d5309f14ee44012759acda59245\": container with ID starting with b915762a27acbf96f1aad344c6de71c18cb55d5309f14ee44012759acda59245 not found: ID does not exist" containerID="b915762a27acbf96f1aad344c6de71c18cb55d5309f14ee44012759acda59245" Dec 03 00:20:24 crc kubenswrapper[4953]: I1203 00:20:24.125734 4953 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b915762a27acbf96f1aad344c6de71c18cb55d5309f14ee44012759acda59245"} err="failed to get container status \"b915762a27acbf96f1aad344c6de71c18cb55d5309f14ee44012759acda59245\": rpc error: code = NotFound desc = could not find container \"b915762a27acbf96f1aad344c6de71c18cb55d5309f14ee44012759acda59245\": container with ID starting with b915762a27acbf96f1aad344c6de71c18cb55d5309f14ee44012759acda59245 not found: ID does not exist" Dec 03 00:20:25 crc kubenswrapper[4953]: I1203 00:20:25.358729 4953 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c2b69c71-8e67-461d-9ddf-cfb439ebb6e1" path="/var/lib/kubelet/pods/c2b69c71-8e67-461d-9ddf-cfb439ebb6e1/volumes" Dec 03 00:20:27 crc kubenswrapper[4953]: I1203 00:20:27.585912 4953 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210kb6ph"] Dec 03 00:20:27 crc kubenswrapper[4953]: E1203 00:20:27.586485 4953 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c2b69c71-8e67-461d-9ddf-cfb439ebb6e1" containerName="registry-server" Dec 03 00:20:27 crc kubenswrapper[4953]: I1203 00:20:27.586500 4953 state_mem.go:107] "Deleted CPUSet assignment" podUID="c2b69c71-8e67-461d-9ddf-cfb439ebb6e1" containerName="registry-server" Dec 03 00:20:27 crc kubenswrapper[4953]: E1203 00:20:27.586513 4953 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c2b69c71-8e67-461d-9ddf-cfb439ebb6e1" containerName="extract-utilities" Dec 03 00:20:27 crc kubenswrapper[4953]: I1203 00:20:27.586520 4953 state_mem.go:107] "Deleted CPUSet assignment" podUID="c2b69c71-8e67-461d-9ddf-cfb439ebb6e1" containerName="extract-utilities" Dec 03 00:20:27 crc kubenswrapper[4953]: E1203 00:20:27.586527 4953 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c2b69c71-8e67-461d-9ddf-cfb439ebb6e1" containerName="extract-content" Dec 03 00:20:27 crc kubenswrapper[4953]: I1203 00:20:27.586534 4953 state_mem.go:107] "Deleted CPUSet assignment" podUID="c2b69c71-8e67-461d-9ddf-cfb439ebb6e1" containerName="extract-content" Dec 03 00:20:27 crc kubenswrapper[4953]: I1203 00:20:27.586658 4953 memory_manager.go:354] "RemoveStaleState removing state" podUID="c2b69c71-8e67-461d-9ddf-cfb439ebb6e1" containerName="registry-server" Dec 03 00:20:27 crc kubenswrapper[4953]: I1203 00:20:27.587426 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210kb6ph" Dec 03 00:20:27 crc kubenswrapper[4953]: I1203 00:20:27.590145 4953 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Dec 03 00:20:27 crc kubenswrapper[4953]: I1203 00:20:27.604325 4953 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210kb6ph"] Dec 03 00:20:27 crc kubenswrapper[4953]: I1203 00:20:27.688562 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/474c987a-6ccd-4d96-bfb1-ae4cb96751c6-bundle\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210kb6ph\" (UID: \"474c987a-6ccd-4d96-bfb1-ae4cb96751c6\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210kb6ph" Dec 03 00:20:27 crc kubenswrapper[4953]: I1203 00:20:27.688654 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/474c987a-6ccd-4d96-bfb1-ae4cb96751c6-util\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210kb6ph\" (UID: \"474c987a-6ccd-4d96-bfb1-ae4cb96751c6\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210kb6ph" Dec 03 00:20:27 crc kubenswrapper[4953]: I1203 00:20:27.688692 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nqbc9\" (UniqueName: \"kubernetes.io/projected/474c987a-6ccd-4d96-bfb1-ae4cb96751c6-kube-api-access-nqbc9\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210kb6ph\" (UID: \"474c987a-6ccd-4d96-bfb1-ae4cb96751c6\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210kb6ph" Dec 03 00:20:27 crc kubenswrapper[4953]: I1203 00:20:27.790524 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nqbc9\" (UniqueName: \"kubernetes.io/projected/474c987a-6ccd-4d96-bfb1-ae4cb96751c6-kube-api-access-nqbc9\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210kb6ph\" (UID: \"474c987a-6ccd-4d96-bfb1-ae4cb96751c6\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210kb6ph" Dec 03 00:20:27 crc kubenswrapper[4953]: I1203 00:20:27.790658 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/474c987a-6ccd-4d96-bfb1-ae4cb96751c6-bundle\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210kb6ph\" (UID: \"474c987a-6ccd-4d96-bfb1-ae4cb96751c6\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210kb6ph" Dec 03 00:20:27 crc kubenswrapper[4953]: I1203 00:20:27.790735 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/474c987a-6ccd-4d96-bfb1-ae4cb96751c6-util\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210kb6ph\" (UID: \"474c987a-6ccd-4d96-bfb1-ae4cb96751c6\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210kb6ph" Dec 03 00:20:27 crc kubenswrapper[4953]: I1203 00:20:27.791383 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/474c987a-6ccd-4d96-bfb1-ae4cb96751c6-bundle\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210kb6ph\" (UID: \"474c987a-6ccd-4d96-bfb1-ae4cb96751c6\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210kb6ph" Dec 03 00:20:27 crc kubenswrapper[4953]: I1203 00:20:27.791418 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/474c987a-6ccd-4d96-bfb1-ae4cb96751c6-util\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210kb6ph\" (UID: \"474c987a-6ccd-4d96-bfb1-ae4cb96751c6\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210kb6ph" Dec 03 00:20:27 crc kubenswrapper[4953]: I1203 00:20:27.814795 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nqbc9\" (UniqueName: \"kubernetes.io/projected/474c987a-6ccd-4d96-bfb1-ae4cb96751c6-kube-api-access-nqbc9\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210kb6ph\" (UID: \"474c987a-6ccd-4d96-bfb1-ae4cb96751c6\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210kb6ph" Dec 03 00:20:27 crc kubenswrapper[4953]: I1203 00:20:27.911035 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210kb6ph" Dec 03 00:20:28 crc kubenswrapper[4953]: I1203 00:20:28.147096 4953 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210kb6ph"] Dec 03 00:20:29 crc kubenswrapper[4953]: I1203 00:20:29.105475 4953 generic.go:334] "Generic (PLEG): container finished" podID="474c987a-6ccd-4d96-bfb1-ae4cb96751c6" containerID="8f3283e87483ee022c7103a306268b7c416497e60b91d76e9e5c41a64c54779a" exitCode=0 Dec 03 00:20:29 crc kubenswrapper[4953]: I1203 00:20:29.105575 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210kb6ph" event={"ID":"474c987a-6ccd-4d96-bfb1-ae4cb96751c6","Type":"ContainerDied","Data":"8f3283e87483ee022c7103a306268b7c416497e60b91d76e9e5c41a64c54779a"} Dec 03 00:20:29 crc kubenswrapper[4953]: I1203 00:20:29.105861 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210kb6ph" event={"ID":"474c987a-6ccd-4d96-bfb1-ae4cb96751c6","Type":"ContainerStarted","Data":"1336cfe025ccb0a2f822fe7f8fdde8b11f2daece3365ae96fd86c6cc03206b35"} Dec 03 00:20:29 crc kubenswrapper[4953]: I1203 00:20:29.108190 4953 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 03 00:20:30 crc kubenswrapper[4953]: I1203 00:20:30.338990 4953 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-6wnhj"] Dec 03 00:20:30 crc kubenswrapper[4953]: I1203 00:20:30.343107 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-6wnhj" Dec 03 00:20:30 crc kubenswrapper[4953]: I1203 00:20:30.348288 4953 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-6wnhj"] Dec 03 00:20:30 crc kubenswrapper[4953]: I1203 00:20:30.432634 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/02b1e287-521d-4996-af3f-61e69be97d6f-catalog-content\") pod \"redhat-operators-6wnhj\" (UID: \"02b1e287-521d-4996-af3f-61e69be97d6f\") " pod="openshift-marketplace/redhat-operators-6wnhj" Dec 03 00:20:30 crc kubenswrapper[4953]: I1203 00:20:30.432781 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/02b1e287-521d-4996-af3f-61e69be97d6f-utilities\") pod \"redhat-operators-6wnhj\" (UID: \"02b1e287-521d-4996-af3f-61e69be97d6f\") " pod="openshift-marketplace/redhat-operators-6wnhj" Dec 03 00:20:30 crc kubenswrapper[4953]: I1203 00:20:30.432819 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q6l58\" (UniqueName: \"kubernetes.io/projected/02b1e287-521d-4996-af3f-61e69be97d6f-kube-api-access-q6l58\") pod \"redhat-operators-6wnhj\" (UID: \"02b1e287-521d-4996-af3f-61e69be97d6f\") " pod="openshift-marketplace/redhat-operators-6wnhj" Dec 03 00:20:30 crc kubenswrapper[4953]: I1203 00:20:30.534944 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/02b1e287-521d-4996-af3f-61e69be97d6f-utilities\") pod \"redhat-operators-6wnhj\" (UID: \"02b1e287-521d-4996-af3f-61e69be97d6f\") " pod="openshift-marketplace/redhat-operators-6wnhj" Dec 03 00:20:30 crc kubenswrapper[4953]: I1203 00:20:30.535043 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q6l58\" (UniqueName: \"kubernetes.io/projected/02b1e287-521d-4996-af3f-61e69be97d6f-kube-api-access-q6l58\") pod \"redhat-operators-6wnhj\" (UID: \"02b1e287-521d-4996-af3f-61e69be97d6f\") " pod="openshift-marketplace/redhat-operators-6wnhj" Dec 03 00:20:30 crc kubenswrapper[4953]: I1203 00:20:30.535127 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/02b1e287-521d-4996-af3f-61e69be97d6f-catalog-content\") pod \"redhat-operators-6wnhj\" (UID: \"02b1e287-521d-4996-af3f-61e69be97d6f\") " pod="openshift-marketplace/redhat-operators-6wnhj" Dec 03 00:20:30 crc kubenswrapper[4953]: I1203 00:20:30.535578 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/02b1e287-521d-4996-af3f-61e69be97d6f-utilities\") pod \"redhat-operators-6wnhj\" (UID: \"02b1e287-521d-4996-af3f-61e69be97d6f\") " pod="openshift-marketplace/redhat-operators-6wnhj" Dec 03 00:20:30 crc kubenswrapper[4953]: I1203 00:20:30.535797 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/02b1e287-521d-4996-af3f-61e69be97d6f-catalog-content\") pod \"redhat-operators-6wnhj\" (UID: \"02b1e287-521d-4996-af3f-61e69be97d6f\") " pod="openshift-marketplace/redhat-operators-6wnhj" Dec 03 00:20:30 crc kubenswrapper[4953]: I1203 00:20:30.558240 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q6l58\" (UniqueName: \"kubernetes.io/projected/02b1e287-521d-4996-af3f-61e69be97d6f-kube-api-access-q6l58\") pod \"redhat-operators-6wnhj\" (UID: \"02b1e287-521d-4996-af3f-61e69be97d6f\") " pod="openshift-marketplace/redhat-operators-6wnhj" Dec 03 00:20:30 crc kubenswrapper[4953]: I1203 00:20:30.678946 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-6wnhj" Dec 03 00:20:31 crc kubenswrapper[4953]: I1203 00:20:31.121014 4953 generic.go:334] "Generic (PLEG): container finished" podID="474c987a-6ccd-4d96-bfb1-ae4cb96751c6" containerID="4aebb687b149c411906704c47d6fcd74cb04839346c2814a02453fc011f4ab31" exitCode=0 Dec 03 00:20:31 crc kubenswrapper[4953]: I1203 00:20:31.121476 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210kb6ph" event={"ID":"474c987a-6ccd-4d96-bfb1-ae4cb96751c6","Type":"ContainerDied","Data":"4aebb687b149c411906704c47d6fcd74cb04839346c2814a02453fc011f4ab31"} Dec 03 00:20:31 crc kubenswrapper[4953]: I1203 00:20:31.121746 4953 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-6wnhj"] Dec 03 00:20:31 crc kubenswrapper[4953]: E1203 00:20:31.655252 4953 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod474c987a_6ccd_4d96_bfb1_ae4cb96751c6.slice/crio-fbd3f5b3bf2e4b125255cbe7c2b8a134fcf75d7f820cb76be68c81972b59a67b.scope\": RecentStats: unable to find data in memory cache]" Dec 03 00:20:32 crc kubenswrapper[4953]: I1203 00:20:32.129386 4953 generic.go:334] "Generic (PLEG): container finished" podID="02b1e287-521d-4996-af3f-61e69be97d6f" containerID="85f09f4b828848a53a992332359b9c8653421632013c4fb9cbd14b27387524f2" exitCode=0 Dec 03 00:20:32 crc kubenswrapper[4953]: I1203 00:20:32.129774 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6wnhj" event={"ID":"02b1e287-521d-4996-af3f-61e69be97d6f","Type":"ContainerDied","Data":"85f09f4b828848a53a992332359b9c8653421632013c4fb9cbd14b27387524f2"} Dec 03 00:20:32 crc kubenswrapper[4953]: I1203 00:20:32.129883 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6wnhj" event={"ID":"02b1e287-521d-4996-af3f-61e69be97d6f","Type":"ContainerStarted","Data":"c4f39139e4050b6a80e6342df2da2f16891cda443a42e67b228e1ffeab34a0e9"} Dec 03 00:20:32 crc kubenswrapper[4953]: I1203 00:20:32.137848 4953 generic.go:334] "Generic (PLEG): container finished" podID="474c987a-6ccd-4d96-bfb1-ae4cb96751c6" containerID="fbd3f5b3bf2e4b125255cbe7c2b8a134fcf75d7f820cb76be68c81972b59a67b" exitCode=0 Dec 03 00:20:32 crc kubenswrapper[4953]: I1203 00:20:32.138040 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210kb6ph" event={"ID":"474c987a-6ccd-4d96-bfb1-ae4cb96751c6","Type":"ContainerDied","Data":"fbd3f5b3bf2e4b125255cbe7c2b8a134fcf75d7f820cb76be68c81972b59a67b"} Dec 03 00:20:33 crc kubenswrapper[4953]: I1203 00:20:33.147706 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6wnhj" event={"ID":"02b1e287-521d-4996-af3f-61e69be97d6f","Type":"ContainerStarted","Data":"c2f1ed596114bc6e75909a10414e54a6f114b0f569617071a730447ef3523473"} Dec 03 00:20:33 crc kubenswrapper[4953]: I1203 00:20:33.407288 4953 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210kb6ph" Dec 03 00:20:33 crc kubenswrapper[4953]: I1203 00:20:33.581838 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nqbc9\" (UniqueName: \"kubernetes.io/projected/474c987a-6ccd-4d96-bfb1-ae4cb96751c6-kube-api-access-nqbc9\") pod \"474c987a-6ccd-4d96-bfb1-ae4cb96751c6\" (UID: \"474c987a-6ccd-4d96-bfb1-ae4cb96751c6\") " Dec 03 00:20:33 crc kubenswrapper[4953]: I1203 00:20:33.581915 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/474c987a-6ccd-4d96-bfb1-ae4cb96751c6-util\") pod \"474c987a-6ccd-4d96-bfb1-ae4cb96751c6\" (UID: \"474c987a-6ccd-4d96-bfb1-ae4cb96751c6\") " Dec 03 00:20:33 crc kubenswrapper[4953]: I1203 00:20:33.582069 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/474c987a-6ccd-4d96-bfb1-ae4cb96751c6-bundle\") pod \"474c987a-6ccd-4d96-bfb1-ae4cb96751c6\" (UID: \"474c987a-6ccd-4d96-bfb1-ae4cb96751c6\") " Dec 03 00:20:33 crc kubenswrapper[4953]: I1203 00:20:33.584312 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/474c987a-6ccd-4d96-bfb1-ae4cb96751c6-bundle" (OuterVolumeSpecName: "bundle") pod "474c987a-6ccd-4d96-bfb1-ae4cb96751c6" (UID: "474c987a-6ccd-4d96-bfb1-ae4cb96751c6"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 00:20:33 crc kubenswrapper[4953]: I1203 00:20:33.593843 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/474c987a-6ccd-4d96-bfb1-ae4cb96751c6-kube-api-access-nqbc9" (OuterVolumeSpecName: "kube-api-access-nqbc9") pod "474c987a-6ccd-4d96-bfb1-ae4cb96751c6" (UID: "474c987a-6ccd-4d96-bfb1-ae4cb96751c6"). InnerVolumeSpecName "kube-api-access-nqbc9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:20:33 crc kubenswrapper[4953]: I1203 00:20:33.596812 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/474c987a-6ccd-4d96-bfb1-ae4cb96751c6-util" (OuterVolumeSpecName: "util") pod "474c987a-6ccd-4d96-bfb1-ae4cb96751c6" (UID: "474c987a-6ccd-4d96-bfb1-ae4cb96751c6"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 00:20:33 crc kubenswrapper[4953]: I1203 00:20:33.683679 4953 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nqbc9\" (UniqueName: \"kubernetes.io/projected/474c987a-6ccd-4d96-bfb1-ae4cb96751c6-kube-api-access-nqbc9\") on node \"crc\" DevicePath \"\"" Dec 03 00:20:33 crc kubenswrapper[4953]: I1203 00:20:33.683735 4953 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/474c987a-6ccd-4d96-bfb1-ae4cb96751c6-util\") on node \"crc\" DevicePath \"\"" Dec 03 00:20:33 crc kubenswrapper[4953]: I1203 00:20:33.683749 4953 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/474c987a-6ccd-4d96-bfb1-ae4cb96751c6-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 00:20:34 crc kubenswrapper[4953]: I1203 00:20:34.159308 4953 generic.go:334] "Generic (PLEG): container finished" podID="02b1e287-521d-4996-af3f-61e69be97d6f" containerID="c2f1ed596114bc6e75909a10414e54a6f114b0f569617071a730447ef3523473" exitCode=0 Dec 03 00:20:34 crc kubenswrapper[4953]: I1203 00:20:34.159374 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6wnhj" event={"ID":"02b1e287-521d-4996-af3f-61e69be97d6f","Type":"ContainerDied","Data":"c2f1ed596114bc6e75909a10414e54a6f114b0f569617071a730447ef3523473"} Dec 03 00:20:34 crc kubenswrapper[4953]: I1203 00:20:34.161691 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210kb6ph" event={"ID":"474c987a-6ccd-4d96-bfb1-ae4cb96751c6","Type":"ContainerDied","Data":"1336cfe025ccb0a2f822fe7f8fdde8b11f2daece3365ae96fd86c6cc03206b35"} Dec 03 00:20:34 crc kubenswrapper[4953]: I1203 00:20:34.161728 4953 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1336cfe025ccb0a2f822fe7f8fdde8b11f2daece3365ae96fd86c6cc03206b35" Dec 03 00:20:34 crc kubenswrapper[4953]: I1203 00:20:34.161797 4953 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210kb6ph" Dec 03 00:20:35 crc kubenswrapper[4953]: I1203 00:20:35.172690 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6wnhj" event={"ID":"02b1e287-521d-4996-af3f-61e69be97d6f","Type":"ContainerStarted","Data":"93b5a0c55d334202cb664f5f8a4fa396479cd332d03c99fad6880aa608774afa"} Dec 03 00:20:35 crc kubenswrapper[4953]: I1203 00:20:35.193574 4953 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-6wnhj" podStartSLOduration=2.733968818 podStartE2EDuration="5.193537924s" podCreationTimestamp="2025-12-03 00:20:30 +0000 UTC" firstStartedPulling="2025-12-03 00:20:32.131707091 +0000 UTC m=+821.115987874" lastFinishedPulling="2025-12-03 00:20:34.591276197 +0000 UTC m=+823.575556980" observedRunningTime="2025-12-03 00:20:35.192320204 +0000 UTC m=+824.176600987" watchObservedRunningTime="2025-12-03 00:20:35.193537924 +0000 UTC m=+824.177818707" Dec 03 00:20:36 crc kubenswrapper[4953]: I1203 00:20:36.976219 4953 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/8ed862a309935d5a1c8012df79b93f7fb46e029d4689f7f6ddcb9e7f5ejn8gr"] Dec 03 00:20:36 crc kubenswrapper[4953]: E1203 00:20:36.976713 4953 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="474c987a-6ccd-4d96-bfb1-ae4cb96751c6" containerName="pull" Dec 03 00:20:36 crc kubenswrapper[4953]: I1203 00:20:36.976736 4953 state_mem.go:107] "Deleted CPUSet assignment" podUID="474c987a-6ccd-4d96-bfb1-ae4cb96751c6" containerName="pull" Dec 03 00:20:36 crc kubenswrapper[4953]: E1203 00:20:36.976758 4953 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="474c987a-6ccd-4d96-bfb1-ae4cb96751c6" containerName="extract" Dec 03 00:20:36 crc kubenswrapper[4953]: I1203 00:20:36.976766 4953 state_mem.go:107] "Deleted CPUSet assignment" podUID="474c987a-6ccd-4d96-bfb1-ae4cb96751c6" containerName="extract" Dec 03 00:20:36 crc kubenswrapper[4953]: E1203 00:20:36.976776 4953 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="474c987a-6ccd-4d96-bfb1-ae4cb96751c6" containerName="util" Dec 03 00:20:36 crc kubenswrapper[4953]: I1203 00:20:36.976786 4953 state_mem.go:107] "Deleted CPUSet assignment" podUID="474c987a-6ccd-4d96-bfb1-ae4cb96751c6" containerName="util" Dec 03 00:20:36 crc kubenswrapper[4953]: I1203 00:20:36.976931 4953 memory_manager.go:354] "RemoveStaleState removing state" podUID="474c987a-6ccd-4d96-bfb1-ae4cb96751c6" containerName="extract" Dec 03 00:20:36 crc kubenswrapper[4953]: I1203 00:20:36.978117 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/8ed862a309935d5a1c8012df79b93f7fb46e029d4689f7f6ddcb9e7f5ejn8gr" Dec 03 00:20:36 crc kubenswrapper[4953]: I1203 00:20:36.980700 4953 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Dec 03 00:20:36 crc kubenswrapper[4953]: I1203 00:20:36.991872 4953 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/8ed862a309935d5a1c8012df79b93f7fb46e029d4689f7f6ddcb9e7f5ejn8gr"] Dec 03 00:20:37 crc kubenswrapper[4953]: I1203 00:20:37.049606 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hjxcn\" (UniqueName: \"kubernetes.io/projected/315d7133-df43-414b-a2af-93c2534cf3ba-kube-api-access-hjxcn\") pod \"8ed862a309935d5a1c8012df79b93f7fb46e029d4689f7f6ddcb9e7f5ejn8gr\" (UID: \"315d7133-df43-414b-a2af-93c2534cf3ba\") " pod="openshift-marketplace/8ed862a309935d5a1c8012df79b93f7fb46e029d4689f7f6ddcb9e7f5ejn8gr" Dec 03 00:20:37 crc kubenswrapper[4953]: I1203 00:20:37.049692 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/315d7133-df43-414b-a2af-93c2534cf3ba-util\") pod \"8ed862a309935d5a1c8012df79b93f7fb46e029d4689f7f6ddcb9e7f5ejn8gr\" (UID: \"315d7133-df43-414b-a2af-93c2534cf3ba\") " pod="openshift-marketplace/8ed862a309935d5a1c8012df79b93f7fb46e029d4689f7f6ddcb9e7f5ejn8gr" Dec 03 00:20:37 crc kubenswrapper[4953]: I1203 00:20:37.049745 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/315d7133-df43-414b-a2af-93c2534cf3ba-bundle\") pod \"8ed862a309935d5a1c8012df79b93f7fb46e029d4689f7f6ddcb9e7f5ejn8gr\" (UID: \"315d7133-df43-414b-a2af-93c2534cf3ba\") " pod="openshift-marketplace/8ed862a309935d5a1c8012df79b93f7fb46e029d4689f7f6ddcb9e7f5ejn8gr" Dec 03 00:20:37 crc kubenswrapper[4953]: I1203 00:20:37.151848 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/315d7133-df43-414b-a2af-93c2534cf3ba-util\") pod \"8ed862a309935d5a1c8012df79b93f7fb46e029d4689f7f6ddcb9e7f5ejn8gr\" (UID: \"315d7133-df43-414b-a2af-93c2534cf3ba\") " pod="openshift-marketplace/8ed862a309935d5a1c8012df79b93f7fb46e029d4689f7f6ddcb9e7f5ejn8gr" Dec 03 00:20:37 crc kubenswrapper[4953]: I1203 00:20:37.151948 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/315d7133-df43-414b-a2af-93c2534cf3ba-bundle\") pod \"8ed862a309935d5a1c8012df79b93f7fb46e029d4689f7f6ddcb9e7f5ejn8gr\" (UID: \"315d7133-df43-414b-a2af-93c2534cf3ba\") " pod="openshift-marketplace/8ed862a309935d5a1c8012df79b93f7fb46e029d4689f7f6ddcb9e7f5ejn8gr" Dec 03 00:20:37 crc kubenswrapper[4953]: I1203 00:20:37.152006 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hjxcn\" (UniqueName: \"kubernetes.io/projected/315d7133-df43-414b-a2af-93c2534cf3ba-kube-api-access-hjxcn\") pod \"8ed862a309935d5a1c8012df79b93f7fb46e029d4689f7f6ddcb9e7f5ejn8gr\" (UID: \"315d7133-df43-414b-a2af-93c2534cf3ba\") " pod="openshift-marketplace/8ed862a309935d5a1c8012df79b93f7fb46e029d4689f7f6ddcb9e7f5ejn8gr" Dec 03 00:20:37 crc kubenswrapper[4953]: I1203 00:20:37.152579 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/315d7133-df43-414b-a2af-93c2534cf3ba-util\") pod \"8ed862a309935d5a1c8012df79b93f7fb46e029d4689f7f6ddcb9e7f5ejn8gr\" (UID: \"315d7133-df43-414b-a2af-93c2534cf3ba\") " pod="openshift-marketplace/8ed862a309935d5a1c8012df79b93f7fb46e029d4689f7f6ddcb9e7f5ejn8gr" Dec 03 00:20:37 crc kubenswrapper[4953]: I1203 00:20:37.152717 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/315d7133-df43-414b-a2af-93c2534cf3ba-bundle\") pod \"8ed862a309935d5a1c8012df79b93f7fb46e029d4689f7f6ddcb9e7f5ejn8gr\" (UID: \"315d7133-df43-414b-a2af-93c2534cf3ba\") " pod="openshift-marketplace/8ed862a309935d5a1c8012df79b93f7fb46e029d4689f7f6ddcb9e7f5ejn8gr" Dec 03 00:20:37 crc kubenswrapper[4953]: I1203 00:20:37.174708 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hjxcn\" (UniqueName: \"kubernetes.io/projected/315d7133-df43-414b-a2af-93c2534cf3ba-kube-api-access-hjxcn\") pod \"8ed862a309935d5a1c8012df79b93f7fb46e029d4689f7f6ddcb9e7f5ejn8gr\" (UID: \"315d7133-df43-414b-a2af-93c2534cf3ba\") " pod="openshift-marketplace/8ed862a309935d5a1c8012df79b93f7fb46e029d4689f7f6ddcb9e7f5ejn8gr" Dec 03 00:20:37 crc kubenswrapper[4953]: I1203 00:20:37.293981 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/8ed862a309935d5a1c8012df79b93f7fb46e029d4689f7f6ddcb9e7f5ejn8gr" Dec 03 00:20:37 crc kubenswrapper[4953]: I1203 00:20:37.509552 4953 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/8ed862a309935d5a1c8012df79b93f7fb46e029d4689f7f6ddcb9e7f5ejn8gr"] Dec 03 00:20:37 crc kubenswrapper[4953]: I1203 00:20:37.964943 4953 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fj4nbn"] Dec 03 00:20:37 crc kubenswrapper[4953]: I1203 00:20:37.966268 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fj4nbn" Dec 03 00:20:37 crc kubenswrapper[4953]: I1203 00:20:37.981312 4953 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fj4nbn"] Dec 03 00:20:38 crc kubenswrapper[4953]: I1203 00:20:38.066314 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/76a42fdc-9500-45c7-b826-5697261a08b0-bundle\") pod \"6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fj4nbn\" (UID: \"76a42fdc-9500-45c7-b826-5697261a08b0\") " pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fj4nbn" Dec 03 00:20:38 crc kubenswrapper[4953]: I1203 00:20:38.066365 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dzn6k\" (UniqueName: \"kubernetes.io/projected/76a42fdc-9500-45c7-b826-5697261a08b0-kube-api-access-dzn6k\") pod \"6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fj4nbn\" (UID: \"76a42fdc-9500-45c7-b826-5697261a08b0\") " pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fj4nbn" Dec 03 00:20:38 crc kubenswrapper[4953]: I1203 00:20:38.066415 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/76a42fdc-9500-45c7-b826-5697261a08b0-util\") pod \"6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fj4nbn\" (UID: \"76a42fdc-9500-45c7-b826-5697261a08b0\") " pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fj4nbn" Dec 03 00:20:38 crc kubenswrapper[4953]: I1203 00:20:38.168106 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/76a42fdc-9500-45c7-b826-5697261a08b0-util\") pod \"6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fj4nbn\" (UID: \"76a42fdc-9500-45c7-b826-5697261a08b0\") " pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fj4nbn" Dec 03 00:20:38 crc kubenswrapper[4953]: I1203 00:20:38.168244 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/76a42fdc-9500-45c7-b826-5697261a08b0-bundle\") pod \"6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fj4nbn\" (UID: \"76a42fdc-9500-45c7-b826-5697261a08b0\") " pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fj4nbn" Dec 03 00:20:38 crc kubenswrapper[4953]: I1203 00:20:38.168282 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dzn6k\" (UniqueName: \"kubernetes.io/projected/76a42fdc-9500-45c7-b826-5697261a08b0-kube-api-access-dzn6k\") pod \"6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fj4nbn\" (UID: \"76a42fdc-9500-45c7-b826-5697261a08b0\") " pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fj4nbn" Dec 03 00:20:38 crc kubenswrapper[4953]: I1203 00:20:38.169087 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/76a42fdc-9500-45c7-b826-5697261a08b0-util\") pod \"6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fj4nbn\" (UID: \"76a42fdc-9500-45c7-b826-5697261a08b0\") " pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fj4nbn" Dec 03 00:20:38 crc kubenswrapper[4953]: I1203 00:20:38.169126 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/76a42fdc-9500-45c7-b826-5697261a08b0-bundle\") pod \"6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fj4nbn\" (UID: \"76a42fdc-9500-45c7-b826-5697261a08b0\") " pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fj4nbn" Dec 03 00:20:38 crc kubenswrapper[4953]: I1203 00:20:38.201849 4953 generic.go:334] "Generic (PLEG): container finished" podID="315d7133-df43-414b-a2af-93c2534cf3ba" containerID="efa78f78add5eacaf764eb1590b36658a511a73d2bef308b5935f9ee7b6431ea" exitCode=0 Dec 03 00:20:38 crc kubenswrapper[4953]: I1203 00:20:38.201957 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/8ed862a309935d5a1c8012df79b93f7fb46e029d4689f7f6ddcb9e7f5ejn8gr" event={"ID":"315d7133-df43-414b-a2af-93c2534cf3ba","Type":"ContainerDied","Data":"efa78f78add5eacaf764eb1590b36658a511a73d2bef308b5935f9ee7b6431ea"} Dec 03 00:20:38 crc kubenswrapper[4953]: I1203 00:20:38.202098 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/8ed862a309935d5a1c8012df79b93f7fb46e029d4689f7f6ddcb9e7f5ejn8gr" event={"ID":"315d7133-df43-414b-a2af-93c2534cf3ba","Type":"ContainerStarted","Data":"036a276ecf6108dd2abfec715ab8884facfc7aff46f0e98dd23f7611dc1c6d3d"} Dec 03 00:20:38 crc kubenswrapper[4953]: I1203 00:20:38.203055 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dzn6k\" (UniqueName: \"kubernetes.io/projected/76a42fdc-9500-45c7-b826-5697261a08b0-kube-api-access-dzn6k\") pod \"6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fj4nbn\" (UID: \"76a42fdc-9500-45c7-b826-5697261a08b0\") " pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fj4nbn" Dec 03 00:20:38 crc kubenswrapper[4953]: I1203 00:20:38.283538 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fj4nbn" Dec 03 00:20:38 crc kubenswrapper[4953]: I1203 00:20:38.517758 4953 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fj4nbn"] Dec 03 00:20:39 crc kubenswrapper[4953]: I1203 00:20:39.209713 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fj4nbn" event={"ID":"76a42fdc-9500-45c7-b826-5697261a08b0","Type":"ContainerStarted","Data":"069b7ebd2ec03040db6718dfb92baedd1699641960c3ce02db9fea9d550ff8e8"} Dec 03 00:20:40 crc kubenswrapper[4953]: I1203 00:20:40.679677 4953 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-6wnhj" Dec 03 00:20:40 crc kubenswrapper[4953]: I1203 00:20:40.734331 4953 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-6wnhj" Dec 03 00:20:40 crc kubenswrapper[4953]: I1203 00:20:40.793306 4953 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-6wnhj" Dec 03 00:20:41 crc kubenswrapper[4953]: I1203 00:20:41.342477 4953 generic.go:334] "Generic (PLEG): container finished" podID="76a42fdc-9500-45c7-b826-5697261a08b0" containerID="e8bb2c031d6ec5f72eb9d9a5c5f05a58d058211888f70dfa3da68ab9c5198d23" exitCode=0 Dec 03 00:20:41 crc kubenswrapper[4953]: I1203 00:20:41.344460 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fj4nbn" event={"ID":"76a42fdc-9500-45c7-b826-5697261a08b0","Type":"ContainerDied","Data":"e8bb2c031d6ec5f72eb9d9a5c5f05a58d058211888f70dfa3da68ab9c5198d23"} Dec 03 00:20:41 crc kubenswrapper[4953]: I1203 00:20:41.454852 4953 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-6wnhj" Dec 03 00:20:41 crc kubenswrapper[4953]: I1203 00:20:41.896212 4953 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-gglz9"] Dec 03 00:20:41 crc kubenswrapper[4953]: I1203 00:20:41.897851 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-gglz9" Dec 03 00:20:42 crc kubenswrapper[4953]: I1203 00:20:42.027624 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/37d22f65-99db-40b6-85ce-a22f2b448708-utilities\") pod \"certified-operators-gglz9\" (UID: \"37d22f65-99db-40b6-85ce-a22f2b448708\") " pod="openshift-marketplace/certified-operators-gglz9" Dec 03 00:20:42 crc kubenswrapper[4953]: I1203 00:20:42.027959 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ljp92\" (UniqueName: \"kubernetes.io/projected/37d22f65-99db-40b6-85ce-a22f2b448708-kube-api-access-ljp92\") pod \"certified-operators-gglz9\" (UID: \"37d22f65-99db-40b6-85ce-a22f2b448708\") " pod="openshift-marketplace/certified-operators-gglz9" Dec 03 00:20:42 crc kubenswrapper[4953]: I1203 00:20:42.028103 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/37d22f65-99db-40b6-85ce-a22f2b448708-catalog-content\") pod \"certified-operators-gglz9\" (UID: \"37d22f65-99db-40b6-85ce-a22f2b448708\") " pod="openshift-marketplace/certified-operators-gglz9" Dec 03 00:20:42 crc kubenswrapper[4953]: I1203 00:20:42.161649 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/37d22f65-99db-40b6-85ce-a22f2b448708-utilities\") pod \"certified-operators-gglz9\" (UID: \"37d22f65-99db-40b6-85ce-a22f2b448708\") " pod="openshift-marketplace/certified-operators-gglz9" Dec 03 00:20:42 crc kubenswrapper[4953]: I1203 00:20:42.162120 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ljp92\" (UniqueName: \"kubernetes.io/projected/37d22f65-99db-40b6-85ce-a22f2b448708-kube-api-access-ljp92\") pod \"certified-operators-gglz9\" (UID: \"37d22f65-99db-40b6-85ce-a22f2b448708\") " pod="openshift-marketplace/certified-operators-gglz9" Dec 03 00:20:42 crc kubenswrapper[4953]: I1203 00:20:42.170160 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/37d22f65-99db-40b6-85ce-a22f2b448708-catalog-content\") pod \"certified-operators-gglz9\" (UID: \"37d22f65-99db-40b6-85ce-a22f2b448708\") " pod="openshift-marketplace/certified-operators-gglz9" Dec 03 00:20:42 crc kubenswrapper[4953]: I1203 00:20:42.171179 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/37d22f65-99db-40b6-85ce-a22f2b448708-catalog-content\") pod \"certified-operators-gglz9\" (UID: \"37d22f65-99db-40b6-85ce-a22f2b448708\") " pod="openshift-marketplace/certified-operators-gglz9" Dec 03 00:20:42 crc kubenswrapper[4953]: I1203 00:20:42.162776 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/37d22f65-99db-40b6-85ce-a22f2b448708-utilities\") pod \"certified-operators-gglz9\" (UID: \"37d22f65-99db-40b6-85ce-a22f2b448708\") " pod="openshift-marketplace/certified-operators-gglz9" Dec 03 00:20:42 crc kubenswrapper[4953]: I1203 00:20:42.507606 4953 generic.go:334] "Generic (PLEG): container finished" podID="315d7133-df43-414b-a2af-93c2534cf3ba" containerID="d5adbd90b4470acbd95396c6a292fe054ae1bbe9aeabf02758ff7ea460a48ef0" exitCode=0 Dec 03 00:20:42 crc kubenswrapper[4953]: I1203 00:20:42.509295 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/8ed862a309935d5a1c8012df79b93f7fb46e029d4689f7f6ddcb9e7f5ejn8gr" event={"ID":"315d7133-df43-414b-a2af-93c2534cf3ba","Type":"ContainerDied","Data":"d5adbd90b4470acbd95396c6a292fe054ae1bbe9aeabf02758ff7ea460a48ef0"} Dec 03 00:20:42 crc kubenswrapper[4953]: I1203 00:20:42.912802 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ljp92\" (UniqueName: \"kubernetes.io/projected/37d22f65-99db-40b6-85ce-a22f2b448708-kube-api-access-ljp92\") pod \"certified-operators-gglz9\" (UID: \"37d22f65-99db-40b6-85ce-a22f2b448708\") " pod="openshift-marketplace/certified-operators-gglz9" Dec 03 00:20:42 crc kubenswrapper[4953]: I1203 00:20:42.927119 4953 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-gglz9"] Dec 03 00:20:43 crc kubenswrapper[4953]: I1203 00:20:43.077958 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-gglz9" Dec 03 00:20:44 crc kubenswrapper[4953]: I1203 00:20:44.246312 4953 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931ahqk5h"] Dec 03 00:20:44 crc kubenswrapper[4953]: I1203 00:20:44.275514 4953 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931ahqk5h"] Dec 03 00:20:44 crc kubenswrapper[4953]: I1203 00:20:44.275732 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931ahqk5h" Dec 03 00:20:44 crc kubenswrapper[4953]: I1203 00:20:44.326299 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/8bc0877b-282d-457a-a46a-c3202a31e2fb-util\") pod \"1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931ahqk5h\" (UID: \"8bc0877b-282d-457a-a46a-c3202a31e2fb\") " pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931ahqk5h" Dec 03 00:20:44 crc kubenswrapper[4953]: I1203 00:20:44.326493 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-58gm2\" (UniqueName: \"kubernetes.io/projected/8bc0877b-282d-457a-a46a-c3202a31e2fb-kube-api-access-58gm2\") pod \"1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931ahqk5h\" (UID: \"8bc0877b-282d-457a-a46a-c3202a31e2fb\") " pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931ahqk5h" Dec 03 00:20:44 crc kubenswrapper[4953]: I1203 00:20:44.326601 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/8bc0877b-282d-457a-a46a-c3202a31e2fb-bundle\") pod \"1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931ahqk5h\" (UID: \"8bc0877b-282d-457a-a46a-c3202a31e2fb\") " pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931ahqk5h" Dec 03 00:20:44 crc kubenswrapper[4953]: I1203 00:20:44.425550 4953 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-gglz9"] Dec 03 00:20:44 crc kubenswrapper[4953]: I1203 00:20:44.434222 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/8bc0877b-282d-457a-a46a-c3202a31e2fb-bundle\") pod \"1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931ahqk5h\" (UID: \"8bc0877b-282d-457a-a46a-c3202a31e2fb\") " pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931ahqk5h" Dec 03 00:20:44 crc kubenswrapper[4953]: I1203 00:20:44.434334 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/8bc0877b-282d-457a-a46a-c3202a31e2fb-util\") pod \"1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931ahqk5h\" (UID: \"8bc0877b-282d-457a-a46a-c3202a31e2fb\") " pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931ahqk5h" Dec 03 00:20:44 crc kubenswrapper[4953]: I1203 00:20:44.434398 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-58gm2\" (UniqueName: \"kubernetes.io/projected/8bc0877b-282d-457a-a46a-c3202a31e2fb-kube-api-access-58gm2\") pod \"1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931ahqk5h\" (UID: \"8bc0877b-282d-457a-a46a-c3202a31e2fb\") " pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931ahqk5h" Dec 03 00:20:44 crc kubenswrapper[4953]: I1203 00:20:44.435165 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/8bc0877b-282d-457a-a46a-c3202a31e2fb-bundle\") pod \"1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931ahqk5h\" (UID: \"8bc0877b-282d-457a-a46a-c3202a31e2fb\") " pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931ahqk5h" Dec 03 00:20:44 crc kubenswrapper[4953]: I1203 00:20:44.435258 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/8bc0877b-282d-457a-a46a-c3202a31e2fb-util\") pod \"1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931ahqk5h\" (UID: \"8bc0877b-282d-457a-a46a-c3202a31e2fb\") " pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931ahqk5h" Dec 03 00:20:44 crc kubenswrapper[4953]: I1203 00:20:44.464600 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-58gm2\" (UniqueName: \"kubernetes.io/projected/8bc0877b-282d-457a-a46a-c3202a31e2fb-kube-api-access-58gm2\") pod \"1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931ahqk5h\" (UID: \"8bc0877b-282d-457a-a46a-c3202a31e2fb\") " pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931ahqk5h" Dec 03 00:20:44 crc kubenswrapper[4953]: I1203 00:20:44.566282 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-gglz9" event={"ID":"37d22f65-99db-40b6-85ce-a22f2b448708","Type":"ContainerStarted","Data":"53a10dec83122a9e74fac5ca93d650561cd19cab7f08d9cb75f8c200d59a2002"} Dec 03 00:20:44 crc kubenswrapper[4953]: I1203 00:20:44.588253 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/8ed862a309935d5a1c8012df79b93f7fb46e029d4689f7f6ddcb9e7f5ejn8gr" event={"ID":"315d7133-df43-414b-a2af-93c2534cf3ba","Type":"ContainerStarted","Data":"6e99e0889f43eb4b7f0cd0e32103f0be0f5da9c559c647c76cd17c04a99a842b"} Dec 03 00:20:44 crc kubenswrapper[4953]: I1203 00:20:44.642728 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931ahqk5h" Dec 03 00:20:44 crc kubenswrapper[4953]: I1203 00:20:44.953833 4953 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/8ed862a309935d5a1c8012df79b93f7fb46e029d4689f7f6ddcb9e7f5ejn8gr" podStartSLOduration=6.270687897 podStartE2EDuration="8.953796002s" podCreationTimestamp="2025-12-03 00:20:36 +0000 UTC" firstStartedPulling="2025-12-03 00:20:38.203906653 +0000 UTC m=+827.188187436" lastFinishedPulling="2025-12-03 00:20:40.887014758 +0000 UTC m=+829.871295541" observedRunningTime="2025-12-03 00:20:44.631305838 +0000 UTC m=+833.615586621" watchObservedRunningTime="2025-12-03 00:20:44.953796002 +0000 UTC m=+833.938076885" Dec 03 00:20:44 crc kubenswrapper[4953]: I1203 00:20:44.958999 4953 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-6wnhj"] Dec 03 00:20:44 crc kubenswrapper[4953]: I1203 00:20:44.959393 4953 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-6wnhj" podUID="02b1e287-521d-4996-af3f-61e69be97d6f" containerName="registry-server" containerID="cri-o://93b5a0c55d334202cb664f5f8a4fa396479cd332d03c99fad6880aa608774afa" gracePeriod=2 Dec 03 00:20:45 crc kubenswrapper[4953]: I1203 00:20:45.498249 4953 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931ahqk5h"] Dec 03 00:20:45 crc kubenswrapper[4953]: I1203 00:20:45.597992 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fj4nbn" event={"ID":"76a42fdc-9500-45c7-b826-5697261a08b0","Type":"ContainerStarted","Data":"e282abfb89b0025d36973c8d218d27a7031f2c1096e17129ad558198348e1e9e"} Dec 03 00:20:45 crc kubenswrapper[4953]: I1203 00:20:45.604534 4953 generic.go:334] "Generic (PLEG): container finished" podID="02b1e287-521d-4996-af3f-61e69be97d6f" containerID="93b5a0c55d334202cb664f5f8a4fa396479cd332d03c99fad6880aa608774afa" exitCode=0 Dec 03 00:20:45 crc kubenswrapper[4953]: I1203 00:20:45.604622 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6wnhj" event={"ID":"02b1e287-521d-4996-af3f-61e69be97d6f","Type":"ContainerDied","Data":"93b5a0c55d334202cb664f5f8a4fa396479cd332d03c99fad6880aa608774afa"} Dec 03 00:20:45 crc kubenswrapper[4953]: I1203 00:20:45.607382 4953 generic.go:334] "Generic (PLEG): container finished" podID="315d7133-df43-414b-a2af-93c2534cf3ba" containerID="6e99e0889f43eb4b7f0cd0e32103f0be0f5da9c559c647c76cd17c04a99a842b" exitCode=0 Dec 03 00:20:45 crc kubenswrapper[4953]: I1203 00:20:45.607477 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/8ed862a309935d5a1c8012df79b93f7fb46e029d4689f7f6ddcb9e7f5ejn8gr" event={"ID":"315d7133-df43-414b-a2af-93c2534cf3ba","Type":"ContainerDied","Data":"6e99e0889f43eb4b7f0cd0e32103f0be0f5da9c559c647c76cd17c04a99a842b"} Dec 03 00:20:45 crc kubenswrapper[4953]: I1203 00:20:45.608483 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931ahqk5h" event={"ID":"8bc0877b-282d-457a-a46a-c3202a31e2fb","Type":"ContainerStarted","Data":"3bb493a46a0fcd601b96c0285efb54caa4e7a7cf5555f8d77a60711313c87ba8"} Dec 03 00:20:45 crc kubenswrapper[4953]: I1203 00:20:45.609585 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-gglz9" event={"ID":"37d22f65-99db-40b6-85ce-a22f2b448708","Type":"ContainerStarted","Data":"15e4958e6f63fcd52549111ef31ed1df48c171486f7a5f1de8354a51d90ea9af"} Dec 03 00:20:46 crc kubenswrapper[4953]: I1203 00:20:46.628185 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931ahqk5h" event={"ID":"8bc0877b-282d-457a-a46a-c3202a31e2fb","Type":"ContainerStarted","Data":"b74b9dc7a9e859c5d9bbf37d6cba2a4754f80b7a087ff7b600df198762dbf5a9"} Dec 03 00:20:46 crc kubenswrapper[4953]: I1203 00:20:46.640284 4953 generic.go:334] "Generic (PLEG): container finished" podID="37d22f65-99db-40b6-85ce-a22f2b448708" containerID="15e4958e6f63fcd52549111ef31ed1df48c171486f7a5f1de8354a51d90ea9af" exitCode=0 Dec 03 00:20:46 crc kubenswrapper[4953]: I1203 00:20:46.640421 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-gglz9" event={"ID":"37d22f65-99db-40b6-85ce-a22f2b448708","Type":"ContainerDied","Data":"15e4958e6f63fcd52549111ef31ed1df48c171486f7a5f1de8354a51d90ea9af"} Dec 03 00:20:46 crc kubenswrapper[4953]: I1203 00:20:46.667832 4953 generic.go:334] "Generic (PLEG): container finished" podID="76a42fdc-9500-45c7-b826-5697261a08b0" containerID="e282abfb89b0025d36973c8d218d27a7031f2c1096e17129ad558198348e1e9e" exitCode=0 Dec 03 00:20:46 crc kubenswrapper[4953]: I1203 00:20:46.668077 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fj4nbn" event={"ID":"76a42fdc-9500-45c7-b826-5697261a08b0","Type":"ContainerDied","Data":"e282abfb89b0025d36973c8d218d27a7031f2c1096e17129ad558198348e1e9e"} Dec 03 00:20:46 crc kubenswrapper[4953]: I1203 00:20:46.680824 4953 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-6wnhj" Dec 03 00:20:46 crc kubenswrapper[4953]: I1203 00:20:46.708681 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6wnhj" event={"ID":"02b1e287-521d-4996-af3f-61e69be97d6f","Type":"ContainerDied","Data":"c4f39139e4050b6a80e6342df2da2f16891cda443a42e67b228e1ffeab34a0e9"} Dec 03 00:20:46 crc kubenswrapper[4953]: I1203 00:20:46.708755 4953 scope.go:117] "RemoveContainer" containerID="93b5a0c55d334202cb664f5f8a4fa396479cd332d03c99fad6880aa608774afa" Dec 03 00:20:46 crc kubenswrapper[4953]: I1203 00:20:46.830569 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q6l58\" (UniqueName: \"kubernetes.io/projected/02b1e287-521d-4996-af3f-61e69be97d6f-kube-api-access-q6l58\") pod \"02b1e287-521d-4996-af3f-61e69be97d6f\" (UID: \"02b1e287-521d-4996-af3f-61e69be97d6f\") " Dec 03 00:20:46 crc kubenswrapper[4953]: I1203 00:20:46.830714 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/02b1e287-521d-4996-af3f-61e69be97d6f-catalog-content\") pod \"02b1e287-521d-4996-af3f-61e69be97d6f\" (UID: \"02b1e287-521d-4996-af3f-61e69be97d6f\") " Dec 03 00:20:46 crc kubenswrapper[4953]: I1203 00:20:46.830773 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/02b1e287-521d-4996-af3f-61e69be97d6f-utilities\") pod \"02b1e287-521d-4996-af3f-61e69be97d6f\" (UID: \"02b1e287-521d-4996-af3f-61e69be97d6f\") " Dec 03 00:20:46 crc kubenswrapper[4953]: I1203 00:20:46.834381 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/02b1e287-521d-4996-af3f-61e69be97d6f-utilities" (OuterVolumeSpecName: "utilities") pod "02b1e287-521d-4996-af3f-61e69be97d6f" (UID: "02b1e287-521d-4996-af3f-61e69be97d6f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 00:20:46 crc kubenswrapper[4953]: I1203 00:20:46.865406 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/02b1e287-521d-4996-af3f-61e69be97d6f-kube-api-access-q6l58" (OuterVolumeSpecName: "kube-api-access-q6l58") pod "02b1e287-521d-4996-af3f-61e69be97d6f" (UID: "02b1e287-521d-4996-af3f-61e69be97d6f"). InnerVolumeSpecName "kube-api-access-q6l58". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:20:46 crc kubenswrapper[4953]: I1203 00:20:46.934323 4953 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q6l58\" (UniqueName: \"kubernetes.io/projected/02b1e287-521d-4996-af3f-61e69be97d6f-kube-api-access-q6l58\") on node \"crc\" DevicePath \"\"" Dec 03 00:20:46 crc kubenswrapper[4953]: I1203 00:20:46.936258 4953 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/02b1e287-521d-4996-af3f-61e69be97d6f-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 00:20:46 crc kubenswrapper[4953]: I1203 00:20:46.995420 4953 scope.go:117] "RemoveContainer" containerID="c2f1ed596114bc6e75909a10414e54a6f114b0f569617071a730447ef3523473" Dec 03 00:20:47 crc kubenswrapper[4953]: I1203 00:20:47.031693 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/02b1e287-521d-4996-af3f-61e69be97d6f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "02b1e287-521d-4996-af3f-61e69be97d6f" (UID: "02b1e287-521d-4996-af3f-61e69be97d6f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 00:20:47 crc kubenswrapper[4953]: I1203 00:20:47.037889 4953 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/02b1e287-521d-4996-af3f-61e69be97d6f-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 00:20:47 crc kubenswrapper[4953]: I1203 00:20:47.067693 4953 scope.go:117] "RemoveContainer" containerID="85f09f4b828848a53a992332359b9c8653421632013c4fb9cbd14b27387524f2" Dec 03 00:20:47 crc kubenswrapper[4953]: I1203 00:20:47.303281 4953 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/8ed862a309935d5a1c8012df79b93f7fb46e029d4689f7f6ddcb9e7f5ejn8gr" Dec 03 00:20:47 crc kubenswrapper[4953]: I1203 00:20:47.457780 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/315d7133-df43-414b-a2af-93c2534cf3ba-bundle\") pod \"315d7133-df43-414b-a2af-93c2534cf3ba\" (UID: \"315d7133-df43-414b-a2af-93c2534cf3ba\") " Dec 03 00:20:47 crc kubenswrapper[4953]: I1203 00:20:47.459386 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hjxcn\" (UniqueName: \"kubernetes.io/projected/315d7133-df43-414b-a2af-93c2534cf3ba-kube-api-access-hjxcn\") pod \"315d7133-df43-414b-a2af-93c2534cf3ba\" (UID: \"315d7133-df43-414b-a2af-93c2534cf3ba\") " Dec 03 00:20:47 crc kubenswrapper[4953]: I1203 00:20:47.459512 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/315d7133-df43-414b-a2af-93c2534cf3ba-util\") pod \"315d7133-df43-414b-a2af-93c2534cf3ba\" (UID: \"315d7133-df43-414b-a2af-93c2534cf3ba\") " Dec 03 00:20:47 crc kubenswrapper[4953]: I1203 00:20:47.458848 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/315d7133-df43-414b-a2af-93c2534cf3ba-bundle" (OuterVolumeSpecName: "bundle") pod "315d7133-df43-414b-a2af-93c2534cf3ba" (UID: "315d7133-df43-414b-a2af-93c2534cf3ba"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 00:20:47 crc kubenswrapper[4953]: I1203 00:20:47.465667 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/315d7133-df43-414b-a2af-93c2534cf3ba-kube-api-access-hjxcn" (OuterVolumeSpecName: "kube-api-access-hjxcn") pod "315d7133-df43-414b-a2af-93c2534cf3ba" (UID: "315d7133-df43-414b-a2af-93c2534cf3ba"). InnerVolumeSpecName "kube-api-access-hjxcn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:20:47 crc kubenswrapper[4953]: I1203 00:20:47.471328 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/315d7133-df43-414b-a2af-93c2534cf3ba-util" (OuterVolumeSpecName: "util") pod "315d7133-df43-414b-a2af-93c2534cf3ba" (UID: "315d7133-df43-414b-a2af-93c2534cf3ba"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 00:20:47 crc kubenswrapper[4953]: I1203 00:20:47.561286 4953 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hjxcn\" (UniqueName: \"kubernetes.io/projected/315d7133-df43-414b-a2af-93c2534cf3ba-kube-api-access-hjxcn\") on node \"crc\" DevicePath \"\"" Dec 03 00:20:47 crc kubenswrapper[4953]: I1203 00:20:47.561913 4953 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/315d7133-df43-414b-a2af-93c2534cf3ba-util\") on node \"crc\" DevicePath \"\"" Dec 03 00:20:47 crc kubenswrapper[4953]: I1203 00:20:47.561990 4953 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/315d7133-df43-414b-a2af-93c2534cf3ba-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 00:20:47 crc kubenswrapper[4953]: I1203 00:20:47.721675 4953 generic.go:334] "Generic (PLEG): container finished" podID="76a42fdc-9500-45c7-b826-5697261a08b0" containerID="9865e42a6d759f8dbe373fe39f213f0690ea9859c8d1a7cd7f74cdd30cd3e096" exitCode=0 Dec 03 00:20:47 crc kubenswrapper[4953]: I1203 00:20:47.721769 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fj4nbn" event={"ID":"76a42fdc-9500-45c7-b826-5697261a08b0","Type":"ContainerDied","Data":"9865e42a6d759f8dbe373fe39f213f0690ea9859c8d1a7cd7f74cdd30cd3e096"} Dec 03 00:20:47 crc kubenswrapper[4953]: I1203 00:20:47.723384 4953 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-6wnhj" Dec 03 00:20:47 crc kubenswrapper[4953]: I1203 00:20:47.725742 4953 generic.go:334] "Generic (PLEG): container finished" podID="8bc0877b-282d-457a-a46a-c3202a31e2fb" containerID="b74b9dc7a9e859c5d9bbf37d6cba2a4754f80b7a087ff7b600df198762dbf5a9" exitCode=0 Dec 03 00:20:47 crc kubenswrapper[4953]: I1203 00:20:47.725789 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931ahqk5h" event={"ID":"8bc0877b-282d-457a-a46a-c3202a31e2fb","Type":"ContainerDied","Data":"b74b9dc7a9e859c5d9bbf37d6cba2a4754f80b7a087ff7b600df198762dbf5a9"} Dec 03 00:20:47 crc kubenswrapper[4953]: I1203 00:20:47.734692 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/8ed862a309935d5a1c8012df79b93f7fb46e029d4689f7f6ddcb9e7f5ejn8gr" event={"ID":"315d7133-df43-414b-a2af-93c2534cf3ba","Type":"ContainerDied","Data":"036a276ecf6108dd2abfec715ab8884facfc7aff46f0e98dd23f7611dc1c6d3d"} Dec 03 00:20:47 crc kubenswrapper[4953]: I1203 00:20:47.734967 4953 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="036a276ecf6108dd2abfec715ab8884facfc7aff46f0e98dd23f7611dc1c6d3d" Dec 03 00:20:47 crc kubenswrapper[4953]: I1203 00:20:47.734762 4953 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/8ed862a309935d5a1c8012df79b93f7fb46e029d4689f7f6ddcb9e7f5ejn8gr" Dec 03 00:20:47 crc kubenswrapper[4953]: I1203 00:20:47.831474 4953 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-6wnhj"] Dec 03 00:20:47 crc kubenswrapper[4953]: I1203 00:20:47.834821 4953 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-6wnhj"] Dec 03 00:20:48 crc kubenswrapper[4953]: I1203 00:20:48.745238 4953 generic.go:334] "Generic (PLEG): container finished" podID="37d22f65-99db-40b6-85ce-a22f2b448708" containerID="0a4fd3ee636c0a87cb2d03c8b75bf58b63c94bf55396b1c07ca8d25ae0ee92b1" exitCode=0 Dec 03 00:20:48 crc kubenswrapper[4953]: I1203 00:20:48.746566 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-gglz9" event={"ID":"37d22f65-99db-40b6-85ce-a22f2b448708","Type":"ContainerDied","Data":"0a4fd3ee636c0a87cb2d03c8b75bf58b63c94bf55396b1c07ca8d25ae0ee92b1"} Dec 03 00:20:48 crc kubenswrapper[4953]: I1203 00:20:48.945290 4953 patch_prober.go:28] interesting pod/machine-config-daemon-6dfbn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 00:20:48 crc kubenswrapper[4953]: I1203 00:20:48.945388 4953 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-6dfbn" podUID="53ffb1e9-9a2e-4869-80d7-776f4aa7ef8b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 00:20:48 crc kubenswrapper[4953]: I1203 00:20:48.945487 4953 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-6dfbn" Dec 03 00:20:48 crc kubenswrapper[4953]: I1203 00:20:48.946671 4953 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"553caefdc3000cd9d8eaf08bf4d23cdb2a37135c2cd1053fce9b7f6cb2b25c0b"} pod="openshift-machine-config-operator/machine-config-daemon-6dfbn" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 03 00:20:48 crc kubenswrapper[4953]: I1203 00:20:48.946740 4953 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-6dfbn" podUID="53ffb1e9-9a2e-4869-80d7-776f4aa7ef8b" containerName="machine-config-daemon" containerID="cri-o://553caefdc3000cd9d8eaf08bf4d23cdb2a37135c2cd1053fce9b7f6cb2b25c0b" gracePeriod=600 Dec 03 00:20:48 crc kubenswrapper[4953]: I1203 00:20:48.982346 4953 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/obo-prometheus-operator-668cf9dfbb-kk8d4"] Dec 03 00:20:48 crc kubenswrapper[4953]: E1203 00:20:48.983245 4953 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="315d7133-df43-414b-a2af-93c2534cf3ba" containerName="util" Dec 03 00:20:48 crc kubenswrapper[4953]: I1203 00:20:48.983272 4953 state_mem.go:107] "Deleted CPUSet assignment" podUID="315d7133-df43-414b-a2af-93c2534cf3ba" containerName="util" Dec 03 00:20:48 crc kubenswrapper[4953]: E1203 00:20:48.983292 4953 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="02b1e287-521d-4996-af3f-61e69be97d6f" containerName="extract-utilities" Dec 03 00:20:48 crc kubenswrapper[4953]: I1203 00:20:48.983303 4953 state_mem.go:107] "Deleted CPUSet assignment" podUID="02b1e287-521d-4996-af3f-61e69be97d6f" containerName="extract-utilities" Dec 03 00:20:48 crc kubenswrapper[4953]: E1203 00:20:48.983314 4953 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="315d7133-df43-414b-a2af-93c2534cf3ba" containerName="pull" Dec 03 00:20:48 crc kubenswrapper[4953]: I1203 00:20:48.983322 4953 state_mem.go:107] "Deleted CPUSet assignment" podUID="315d7133-df43-414b-a2af-93c2534cf3ba" containerName="pull" Dec 03 00:20:48 crc kubenswrapper[4953]: E1203 00:20:48.983339 4953 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="02b1e287-521d-4996-af3f-61e69be97d6f" containerName="extract-content" Dec 03 00:20:48 crc kubenswrapper[4953]: I1203 00:20:48.983348 4953 state_mem.go:107] "Deleted CPUSet assignment" podUID="02b1e287-521d-4996-af3f-61e69be97d6f" containerName="extract-content" Dec 03 00:20:48 crc kubenswrapper[4953]: E1203 00:20:48.983358 4953 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="02b1e287-521d-4996-af3f-61e69be97d6f" containerName="registry-server" Dec 03 00:20:48 crc kubenswrapper[4953]: I1203 00:20:48.983368 4953 state_mem.go:107] "Deleted CPUSet assignment" podUID="02b1e287-521d-4996-af3f-61e69be97d6f" containerName="registry-server" Dec 03 00:20:48 crc kubenswrapper[4953]: E1203 00:20:48.983392 4953 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="315d7133-df43-414b-a2af-93c2534cf3ba" containerName="extract" Dec 03 00:20:48 crc kubenswrapper[4953]: I1203 00:20:48.983400 4953 state_mem.go:107] "Deleted CPUSet assignment" podUID="315d7133-df43-414b-a2af-93c2534cf3ba" containerName="extract" Dec 03 00:20:48 crc kubenswrapper[4953]: I1203 00:20:48.983565 4953 memory_manager.go:354] "RemoveStaleState removing state" podUID="315d7133-df43-414b-a2af-93c2534cf3ba" containerName="extract" Dec 03 00:20:48 crc kubenswrapper[4953]: I1203 00:20:48.983583 4953 memory_manager.go:354] "RemoveStaleState removing state" podUID="02b1e287-521d-4996-af3f-61e69be97d6f" containerName="registry-server" Dec 03 00:20:48 crc kubenswrapper[4953]: I1203 00:20:48.984358 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-kk8d4" Dec 03 00:20:48 crc kubenswrapper[4953]: I1203 00:20:48.991723 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operators"/"kube-root-ca.crt" Dec 03 00:20:48 crc kubenswrapper[4953]: I1203 00:20:48.991860 4953 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"obo-prometheus-operator-dockercfg-4zhw7" Dec 03 00:20:48 crc kubenswrapper[4953]: I1203 00:20:48.993591 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operators"/"openshift-service-ca.crt" Dec 03 00:20:49 crc kubenswrapper[4953]: I1203 00:20:48.999063 4953 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-668cf9dfbb-kk8d4"] Dec 03 00:20:49 crc kubenswrapper[4953]: I1203 00:20:49.088540 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-64956\" (UniqueName: \"kubernetes.io/projected/f5d63f17-0a24-453c-a528-e61266f2b091-kube-api-access-64956\") pod \"obo-prometheus-operator-668cf9dfbb-kk8d4\" (UID: \"f5d63f17-0a24-453c-a528-e61266f2b091\") " pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-kk8d4" Dec 03 00:20:49 crc kubenswrapper[4953]: I1203 00:20:49.136393 4953 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-68bbd5c68d-pmttp"] Dec 03 00:20:49 crc kubenswrapper[4953]: I1203 00:20:49.137549 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-68bbd5c68d-pmttp" Dec 03 00:20:49 crc kubenswrapper[4953]: I1203 00:20:49.143624 4953 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"obo-prometheus-operator-admission-webhook-dockercfg-9jtv6" Dec 03 00:20:49 crc kubenswrapper[4953]: I1203 00:20:49.146524 4953 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"obo-prometheus-operator-admission-webhook-service-cert" Dec 03 00:20:49 crc kubenswrapper[4953]: I1203 00:20:49.156950 4953 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-68bbd5c68d-pmttp"] Dec 03 00:20:49 crc kubenswrapper[4953]: I1203 00:20:49.173778 4953 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-68bbd5c68d-2lg5v"] Dec 03 00:20:49 crc kubenswrapper[4953]: I1203 00:20:49.175042 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-68bbd5c68d-2lg5v" Dec 03 00:20:49 crc kubenswrapper[4953]: I1203 00:20:49.189855 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-64956\" (UniqueName: \"kubernetes.io/projected/f5d63f17-0a24-453c-a528-e61266f2b091-kube-api-access-64956\") pod \"obo-prometheus-operator-668cf9dfbb-kk8d4\" (UID: \"f5d63f17-0a24-453c-a528-e61266f2b091\") " pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-kk8d4" Dec 03 00:20:49 crc kubenswrapper[4953]: I1203 00:20:49.195331 4953 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fj4nbn" Dec 03 00:20:49 crc kubenswrapper[4953]: I1203 00:20:49.218381 4953 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-68bbd5c68d-2lg5v"] Dec 03 00:20:49 crc kubenswrapper[4953]: I1203 00:20:49.242332 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-64956\" (UniqueName: \"kubernetes.io/projected/f5d63f17-0a24-453c-a528-e61266f2b091-kube-api-access-64956\") pod \"obo-prometheus-operator-668cf9dfbb-kk8d4\" (UID: \"f5d63f17-0a24-453c-a528-e61266f2b091\") " pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-kk8d4" Dec 03 00:20:49 crc kubenswrapper[4953]: I1203 00:20:49.291900 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/76a42fdc-9500-45c7-b826-5697261a08b0-bundle\") pod \"76a42fdc-9500-45c7-b826-5697261a08b0\" (UID: \"76a42fdc-9500-45c7-b826-5697261a08b0\") " Dec 03 00:20:49 crc kubenswrapper[4953]: I1203 00:20:49.292006 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dzn6k\" (UniqueName: \"kubernetes.io/projected/76a42fdc-9500-45c7-b826-5697261a08b0-kube-api-access-dzn6k\") pod \"76a42fdc-9500-45c7-b826-5697261a08b0\" (UID: \"76a42fdc-9500-45c7-b826-5697261a08b0\") " Dec 03 00:20:49 crc kubenswrapper[4953]: I1203 00:20:49.292081 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/76a42fdc-9500-45c7-b826-5697261a08b0-util\") pod \"76a42fdc-9500-45c7-b826-5697261a08b0\" (UID: \"76a42fdc-9500-45c7-b826-5697261a08b0\") " Dec 03 00:20:49 crc kubenswrapper[4953]: I1203 00:20:49.292349 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/fc13989b-f854-49fd-8901-3b206167aafc-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-68bbd5c68d-pmttp\" (UID: \"fc13989b-f854-49fd-8901-3b206167aafc\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-68bbd5c68d-pmttp" Dec 03 00:20:49 crc kubenswrapper[4953]: I1203 00:20:49.292393 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/fc13989b-f854-49fd-8901-3b206167aafc-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-68bbd5c68d-pmttp\" (UID: \"fc13989b-f854-49fd-8901-3b206167aafc\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-68bbd5c68d-pmttp" Dec 03 00:20:49 crc kubenswrapper[4953]: I1203 00:20:49.292414 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/0ce8b8d2-06e2-4df3-9a3f-d2022263fd97-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-68bbd5c68d-2lg5v\" (UID: \"0ce8b8d2-06e2-4df3-9a3f-d2022263fd97\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-68bbd5c68d-2lg5v" Dec 03 00:20:49 crc kubenswrapper[4953]: I1203 00:20:49.292445 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/0ce8b8d2-06e2-4df3-9a3f-d2022263fd97-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-68bbd5c68d-2lg5v\" (UID: \"0ce8b8d2-06e2-4df3-9a3f-d2022263fd97\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-68bbd5c68d-2lg5v" Dec 03 00:20:49 crc kubenswrapper[4953]: I1203 00:20:49.293206 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/76a42fdc-9500-45c7-b826-5697261a08b0-bundle" (OuterVolumeSpecName: "bundle") pod "76a42fdc-9500-45c7-b826-5697261a08b0" (UID: "76a42fdc-9500-45c7-b826-5697261a08b0"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 00:20:49 crc kubenswrapper[4953]: I1203 00:20:49.296754 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/76a42fdc-9500-45c7-b826-5697261a08b0-kube-api-access-dzn6k" (OuterVolumeSpecName: "kube-api-access-dzn6k") pod "76a42fdc-9500-45c7-b826-5697261a08b0" (UID: "76a42fdc-9500-45c7-b826-5697261a08b0"). InnerVolumeSpecName "kube-api-access-dzn6k". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:20:49 crc kubenswrapper[4953]: I1203 00:20:49.311067 4953 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/observability-operator-d8bb48f5d-64xzt"] Dec 03 00:20:49 crc kubenswrapper[4953]: E1203 00:20:49.311413 4953 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="76a42fdc-9500-45c7-b826-5697261a08b0" containerName="pull" Dec 03 00:20:49 crc kubenswrapper[4953]: I1203 00:20:49.311428 4953 state_mem.go:107] "Deleted CPUSet assignment" podUID="76a42fdc-9500-45c7-b826-5697261a08b0" containerName="pull" Dec 03 00:20:49 crc kubenswrapper[4953]: E1203 00:20:49.311440 4953 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="76a42fdc-9500-45c7-b826-5697261a08b0" containerName="extract" Dec 03 00:20:49 crc kubenswrapper[4953]: I1203 00:20:49.311447 4953 state_mem.go:107] "Deleted CPUSet assignment" podUID="76a42fdc-9500-45c7-b826-5697261a08b0" containerName="extract" Dec 03 00:20:49 crc kubenswrapper[4953]: E1203 00:20:49.311457 4953 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="76a42fdc-9500-45c7-b826-5697261a08b0" containerName="util" Dec 03 00:20:49 crc kubenswrapper[4953]: I1203 00:20:49.311464 4953 state_mem.go:107] "Deleted CPUSet assignment" podUID="76a42fdc-9500-45c7-b826-5697261a08b0" containerName="util" Dec 03 00:20:49 crc kubenswrapper[4953]: I1203 00:20:49.311589 4953 memory_manager.go:354] "RemoveStaleState removing state" podUID="76a42fdc-9500-45c7-b826-5697261a08b0" containerName="extract" Dec 03 00:20:49 crc kubenswrapper[4953]: I1203 00:20:49.312192 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-operator-d8bb48f5d-64xzt" Dec 03 00:20:49 crc kubenswrapper[4953]: I1203 00:20:49.324862 4953 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"observability-operator-tls" Dec 03 00:20:49 crc kubenswrapper[4953]: I1203 00:20:49.328662 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-kk8d4" Dec 03 00:20:49 crc kubenswrapper[4953]: I1203 00:20:49.328749 4953 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"observability-operator-sa-dockercfg-h92dn" Dec 03 00:20:49 crc kubenswrapper[4953]: I1203 00:20:49.331284 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/76a42fdc-9500-45c7-b826-5697261a08b0-util" (OuterVolumeSpecName: "util") pod "76a42fdc-9500-45c7-b826-5697261a08b0" (UID: "76a42fdc-9500-45c7-b826-5697261a08b0"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 00:20:49 crc kubenswrapper[4953]: I1203 00:20:49.393525 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/0ce8b8d2-06e2-4df3-9a3f-d2022263fd97-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-68bbd5c68d-2lg5v\" (UID: \"0ce8b8d2-06e2-4df3-9a3f-d2022263fd97\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-68bbd5c68d-2lg5v" Dec 03 00:20:49 crc kubenswrapper[4953]: I1203 00:20:49.424211 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9dn9l\" (UniqueName: \"kubernetes.io/projected/1df642b8-586a-4bc5-8d9f-9f2cff7c6b80-kube-api-access-9dn9l\") pod \"observability-operator-d8bb48f5d-64xzt\" (UID: \"1df642b8-586a-4bc5-8d9f-9f2cff7c6b80\") " pod="openshift-operators/observability-operator-d8bb48f5d-64xzt" Dec 03 00:20:49 crc kubenswrapper[4953]: I1203 00:20:49.424282 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/fc13989b-f854-49fd-8901-3b206167aafc-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-68bbd5c68d-pmttp\" (UID: \"fc13989b-f854-49fd-8901-3b206167aafc\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-68bbd5c68d-pmttp" Dec 03 00:20:49 crc kubenswrapper[4953]: I1203 00:20:49.424336 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/fc13989b-f854-49fd-8901-3b206167aafc-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-68bbd5c68d-pmttp\" (UID: \"fc13989b-f854-49fd-8901-3b206167aafc\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-68bbd5c68d-pmttp" Dec 03 00:20:49 crc kubenswrapper[4953]: I1203 00:20:49.424361 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/0ce8b8d2-06e2-4df3-9a3f-d2022263fd97-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-68bbd5c68d-2lg5v\" (UID: \"0ce8b8d2-06e2-4df3-9a3f-d2022263fd97\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-68bbd5c68d-2lg5v" Dec 03 00:20:49 crc kubenswrapper[4953]: I1203 00:20:49.424383 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"observability-operator-tls\" (UniqueName: \"kubernetes.io/secret/1df642b8-586a-4bc5-8d9f-9f2cff7c6b80-observability-operator-tls\") pod \"observability-operator-d8bb48f5d-64xzt\" (UID: \"1df642b8-586a-4bc5-8d9f-9f2cff7c6b80\") " pod="openshift-operators/observability-operator-d8bb48f5d-64xzt" Dec 03 00:20:49 crc kubenswrapper[4953]: I1203 00:20:49.424436 4953 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/76a42fdc-9500-45c7-b826-5697261a08b0-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 00:20:49 crc kubenswrapper[4953]: I1203 00:20:49.424454 4953 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dzn6k\" (UniqueName: \"kubernetes.io/projected/76a42fdc-9500-45c7-b826-5697261a08b0-kube-api-access-dzn6k\") on node \"crc\" DevicePath \"\"" Dec 03 00:20:49 crc kubenswrapper[4953]: I1203 00:20:49.424463 4953 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/76a42fdc-9500-45c7-b826-5697261a08b0-util\") on node \"crc\" DevicePath \"\"" Dec 03 00:20:49 crc kubenswrapper[4953]: I1203 00:20:49.407040 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/0ce8b8d2-06e2-4df3-9a3f-d2022263fd97-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-68bbd5c68d-2lg5v\" (UID: \"0ce8b8d2-06e2-4df3-9a3f-d2022263fd97\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-68bbd5c68d-2lg5v" Dec 03 00:20:49 crc kubenswrapper[4953]: I1203 00:20:49.408432 4953 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="02b1e287-521d-4996-af3f-61e69be97d6f" path="/var/lib/kubelet/pods/02b1e287-521d-4996-af3f-61e69be97d6f/volumes" Dec 03 00:20:49 crc kubenswrapper[4953]: I1203 00:20:49.425488 4953 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/observability-operator-d8bb48f5d-64xzt"] Dec 03 00:20:49 crc kubenswrapper[4953]: I1203 00:20:49.432040 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/fc13989b-f854-49fd-8901-3b206167aafc-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-68bbd5c68d-pmttp\" (UID: \"fc13989b-f854-49fd-8901-3b206167aafc\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-68bbd5c68d-pmttp" Dec 03 00:20:49 crc kubenswrapper[4953]: I1203 00:20:49.435644 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/fc13989b-f854-49fd-8901-3b206167aafc-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-68bbd5c68d-pmttp\" (UID: \"fc13989b-f854-49fd-8901-3b206167aafc\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-68bbd5c68d-pmttp" Dec 03 00:20:49 crc kubenswrapper[4953]: I1203 00:20:49.442583 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/0ce8b8d2-06e2-4df3-9a3f-d2022263fd97-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-68bbd5c68d-2lg5v\" (UID: \"0ce8b8d2-06e2-4df3-9a3f-d2022263fd97\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-68bbd5c68d-2lg5v" Dec 03 00:20:49 crc kubenswrapper[4953]: I1203 00:20:49.490384 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-68bbd5c68d-pmttp" Dec 03 00:20:49 crc kubenswrapper[4953]: I1203 00:20:49.505087 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-68bbd5c68d-2lg5v" Dec 03 00:20:49 crc kubenswrapper[4953]: I1203 00:20:49.525556 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9dn9l\" (UniqueName: \"kubernetes.io/projected/1df642b8-586a-4bc5-8d9f-9f2cff7c6b80-kube-api-access-9dn9l\") pod \"observability-operator-d8bb48f5d-64xzt\" (UID: \"1df642b8-586a-4bc5-8d9f-9f2cff7c6b80\") " pod="openshift-operators/observability-operator-d8bb48f5d-64xzt" Dec 03 00:20:49 crc kubenswrapper[4953]: I1203 00:20:49.525656 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"observability-operator-tls\" (UniqueName: \"kubernetes.io/secret/1df642b8-586a-4bc5-8d9f-9f2cff7c6b80-observability-operator-tls\") pod \"observability-operator-d8bb48f5d-64xzt\" (UID: \"1df642b8-586a-4bc5-8d9f-9f2cff7c6b80\") " pod="openshift-operators/observability-operator-d8bb48f5d-64xzt" Dec 03 00:20:49 crc kubenswrapper[4953]: I1203 00:20:49.529776 4953 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/perses-operator-5446b9c989-hqmj8"] Dec 03 00:20:49 crc kubenswrapper[4953]: I1203 00:20:49.537123 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/perses-operator-5446b9c989-hqmj8" Dec 03 00:20:49 crc kubenswrapper[4953]: I1203 00:20:49.538449 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"observability-operator-tls\" (UniqueName: \"kubernetes.io/secret/1df642b8-586a-4bc5-8d9f-9f2cff7c6b80-observability-operator-tls\") pod \"observability-operator-d8bb48f5d-64xzt\" (UID: \"1df642b8-586a-4bc5-8d9f-9f2cff7c6b80\") " pod="openshift-operators/observability-operator-d8bb48f5d-64xzt" Dec 03 00:20:49 crc kubenswrapper[4953]: I1203 00:20:49.540524 4953 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"perses-operator-dockercfg-kqcv6" Dec 03 00:20:49 crc kubenswrapper[4953]: I1203 00:20:49.564890 4953 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/perses-operator-5446b9c989-hqmj8"] Dec 03 00:20:49 crc kubenswrapper[4953]: I1203 00:20:49.610592 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9dn9l\" (UniqueName: \"kubernetes.io/projected/1df642b8-586a-4bc5-8d9f-9f2cff7c6b80-kube-api-access-9dn9l\") pod \"observability-operator-d8bb48f5d-64xzt\" (UID: \"1df642b8-586a-4bc5-8d9f-9f2cff7c6b80\") " pod="openshift-operators/observability-operator-d8bb48f5d-64xzt" Dec 03 00:20:49 crc kubenswrapper[4953]: I1203 00:20:49.628046 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openshift-service-ca\" (UniqueName: \"kubernetes.io/configmap/61476fab-b1b9-424c-ab5d-853f83d6fdfd-openshift-service-ca\") pod \"perses-operator-5446b9c989-hqmj8\" (UID: \"61476fab-b1b9-424c-ab5d-853f83d6fdfd\") " pod="openshift-operators/perses-operator-5446b9c989-hqmj8" Dec 03 00:20:49 crc kubenswrapper[4953]: I1203 00:20:49.628140 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zckhl\" (UniqueName: \"kubernetes.io/projected/61476fab-b1b9-424c-ab5d-853f83d6fdfd-kube-api-access-zckhl\") pod \"perses-operator-5446b9c989-hqmj8\" (UID: \"61476fab-b1b9-424c-ab5d-853f83d6fdfd\") " pod="openshift-operators/perses-operator-5446b9c989-hqmj8" Dec 03 00:20:49 crc kubenswrapper[4953]: I1203 00:20:49.667032 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-operator-d8bb48f5d-64xzt" Dec 03 00:20:49 crc kubenswrapper[4953]: I1203 00:20:49.737042 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zckhl\" (UniqueName: \"kubernetes.io/projected/61476fab-b1b9-424c-ab5d-853f83d6fdfd-kube-api-access-zckhl\") pod \"perses-operator-5446b9c989-hqmj8\" (UID: \"61476fab-b1b9-424c-ab5d-853f83d6fdfd\") " pod="openshift-operators/perses-operator-5446b9c989-hqmj8" Dec 03 00:20:49 crc kubenswrapper[4953]: I1203 00:20:49.737182 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openshift-service-ca\" (UniqueName: \"kubernetes.io/configmap/61476fab-b1b9-424c-ab5d-853f83d6fdfd-openshift-service-ca\") pod \"perses-operator-5446b9c989-hqmj8\" (UID: \"61476fab-b1b9-424c-ab5d-853f83d6fdfd\") " pod="openshift-operators/perses-operator-5446b9c989-hqmj8" Dec 03 00:20:49 crc kubenswrapper[4953]: I1203 00:20:49.738292 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openshift-service-ca\" (UniqueName: \"kubernetes.io/configmap/61476fab-b1b9-424c-ab5d-853f83d6fdfd-openshift-service-ca\") pod \"perses-operator-5446b9c989-hqmj8\" (UID: \"61476fab-b1b9-424c-ab5d-853f83d6fdfd\") " pod="openshift-operators/perses-operator-5446b9c989-hqmj8" Dec 03 00:20:49 crc kubenswrapper[4953]: I1203 00:20:49.801865 4953 generic.go:334] "Generic (PLEG): container finished" podID="53ffb1e9-9a2e-4869-80d7-776f4aa7ef8b" containerID="553caefdc3000cd9d8eaf08bf4d23cdb2a37135c2cd1053fce9b7f6cb2b25c0b" exitCode=0 Dec 03 00:20:49 crc kubenswrapper[4953]: I1203 00:20:49.802022 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-6dfbn" event={"ID":"53ffb1e9-9a2e-4869-80d7-776f4aa7ef8b","Type":"ContainerDied","Data":"553caefdc3000cd9d8eaf08bf4d23cdb2a37135c2cd1053fce9b7f6cb2b25c0b"} Dec 03 00:20:49 crc kubenswrapper[4953]: I1203 00:20:49.802068 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-6dfbn" event={"ID":"53ffb1e9-9a2e-4869-80d7-776f4aa7ef8b","Type":"ContainerStarted","Data":"ede43397367a8a468be56d98b9f73d5afe6a5294be1471b33f69b64147ab3995"} Dec 03 00:20:49 crc kubenswrapper[4953]: I1203 00:20:49.802092 4953 scope.go:117] "RemoveContainer" containerID="257c3b050511c36fa41f2b2785b39713a95149ac43905452ee5690643ac49dc2" Dec 03 00:20:49 crc kubenswrapper[4953]: I1203 00:20:49.829295 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fj4nbn" event={"ID":"76a42fdc-9500-45c7-b826-5697261a08b0","Type":"ContainerDied","Data":"069b7ebd2ec03040db6718dfb92baedd1699641960c3ce02db9fea9d550ff8e8"} Dec 03 00:20:49 crc kubenswrapper[4953]: I1203 00:20:49.829366 4953 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="069b7ebd2ec03040db6718dfb92baedd1699641960c3ce02db9fea9d550ff8e8" Dec 03 00:20:49 crc kubenswrapper[4953]: I1203 00:20:49.829486 4953 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fj4nbn" Dec 03 00:20:49 crc kubenswrapper[4953]: I1203 00:20:49.869173 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zckhl\" (UniqueName: \"kubernetes.io/projected/61476fab-b1b9-424c-ab5d-853f83d6fdfd-kube-api-access-zckhl\") pod \"perses-operator-5446b9c989-hqmj8\" (UID: \"61476fab-b1b9-424c-ab5d-853f83d6fdfd\") " pod="openshift-operators/perses-operator-5446b9c989-hqmj8" Dec 03 00:20:50 crc kubenswrapper[4953]: I1203 00:20:50.165911 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/perses-operator-5446b9c989-hqmj8" Dec 03 00:20:50 crc kubenswrapper[4953]: I1203 00:20:50.283995 4953 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-668cf9dfbb-kk8d4"] Dec 03 00:20:50 crc kubenswrapper[4953]: I1203 00:20:50.660754 4953 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/observability-operator-d8bb48f5d-64xzt"] Dec 03 00:20:50 crc kubenswrapper[4953]: I1203 00:20:50.813342 4953 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-68bbd5c68d-pmttp"] Dec 03 00:20:50 crc kubenswrapper[4953]: I1203 00:20:50.866818 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/observability-operator-d8bb48f5d-64xzt" event={"ID":"1df642b8-586a-4bc5-8d9f-9f2cff7c6b80","Type":"ContainerStarted","Data":"b35f0c33954d7dcccd557c94a2c9d21a59de4dce14d03ee1f75970266f81ed5b"} Dec 03 00:20:50 crc kubenswrapper[4953]: W1203 00:20:50.895028 4953 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podfc13989b_f854_49fd_8901_3b206167aafc.slice/crio-b8c9b5ed9786ee1c0a041dc7d5246422db9a09ebc6db0df723a3e476dccfbb1e WatchSource:0}: Error finding container b8c9b5ed9786ee1c0a041dc7d5246422db9a09ebc6db0df723a3e476dccfbb1e: Status 404 returned error can't find the container with id b8c9b5ed9786ee1c0a041dc7d5246422db9a09ebc6db0df723a3e476dccfbb1e Dec 03 00:20:50 crc kubenswrapper[4953]: I1203 00:20:50.914620 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-gglz9" event={"ID":"37d22f65-99db-40b6-85ce-a22f2b448708","Type":"ContainerStarted","Data":"3090ad29f26f813c98f5d2f7f67d2ff61214d101a507315b4456f25138670998"} Dec 03 00:20:50 crc kubenswrapper[4953]: I1203 00:20:50.923929 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-kk8d4" event={"ID":"f5d63f17-0a24-453c-a528-e61266f2b091","Type":"ContainerStarted","Data":"378dbea25486d132e82a76e62e73d2729409680f96259e30a4ca468b7a5b1f54"} Dec 03 00:20:50 crc kubenswrapper[4953]: I1203 00:20:50.924468 4953 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-68bbd5c68d-2lg5v"] Dec 03 00:20:50 crc kubenswrapper[4953]: I1203 00:20:50.968725 4953 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/perses-operator-5446b9c989-hqmj8"] Dec 03 00:20:50 crc kubenswrapper[4953]: I1203 00:20:50.987479 4953 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-gglz9" podStartSLOduration=7.305480209 podStartE2EDuration="9.987390656s" podCreationTimestamp="2025-12-03 00:20:41 +0000 UTC" firstStartedPulling="2025-12-03 00:20:46.645479853 +0000 UTC m=+835.629760636" lastFinishedPulling="2025-12-03 00:20:49.3273903 +0000 UTC m=+838.311671083" observedRunningTime="2025-12-03 00:20:50.986714709 +0000 UTC m=+839.970995492" watchObservedRunningTime="2025-12-03 00:20:50.987390656 +0000 UTC m=+839.971671439" Dec 03 00:20:51 crc kubenswrapper[4953]: I1203 00:20:51.951396 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-68bbd5c68d-pmttp" event={"ID":"fc13989b-f854-49fd-8901-3b206167aafc","Type":"ContainerStarted","Data":"b8c9b5ed9786ee1c0a041dc7d5246422db9a09ebc6db0df723a3e476dccfbb1e"} Dec 03 00:20:51 crc kubenswrapper[4953]: I1203 00:20:51.976361 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/perses-operator-5446b9c989-hqmj8" event={"ID":"61476fab-b1b9-424c-ab5d-853f83d6fdfd","Type":"ContainerStarted","Data":"15e77fca40c0409921c2218ee3678ec0db0137c5e4d5202ebd11a992f5afbef7"} Dec 03 00:20:51 crc kubenswrapper[4953]: I1203 00:20:51.996811 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-68bbd5c68d-2lg5v" event={"ID":"0ce8b8d2-06e2-4df3-9a3f-d2022263fd97","Type":"ContainerStarted","Data":"da31f78369ca982673ea30ce470cd558232bd39a4faff610f22695b0e102b9c6"} Dec 03 00:20:53 crc kubenswrapper[4953]: I1203 00:20:53.080800 4953 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-gglz9" Dec 03 00:20:53 crc kubenswrapper[4953]: I1203 00:20:53.082914 4953 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-gglz9" Dec 03 00:20:53 crc kubenswrapper[4953]: I1203 00:20:53.210943 4953 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-gglz9" Dec 03 00:20:54 crc kubenswrapper[4953]: I1203 00:20:54.531418 4953 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-gglz9" Dec 03 00:20:57 crc kubenswrapper[4953]: I1203 00:20:57.675741 4953 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-gglz9"] Dec 03 00:20:57 crc kubenswrapper[4953]: I1203 00:20:57.676617 4953 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-gglz9" podUID="37d22f65-99db-40b6-85ce-a22f2b448708" containerName="registry-server" containerID="cri-o://3090ad29f26f813c98f5d2f7f67d2ff61214d101a507315b4456f25138670998" gracePeriod=2 Dec 03 00:20:58 crc kubenswrapper[4953]: I1203 00:20:58.035502 4953 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["service-telemetry/elastic-operator-866854d65-qhl5q"] Dec 03 00:20:58 crc kubenswrapper[4953]: I1203 00:20:58.036536 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/elastic-operator-866854d65-qhl5q" Dec 03 00:20:58 crc kubenswrapper[4953]: I1203 00:20:58.040410 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"kube-root-ca.crt" Dec 03 00:20:58 crc kubenswrapper[4953]: I1203 00:20:58.040777 4953 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"elastic-operator-service-cert" Dec 03 00:20:58 crc kubenswrapper[4953]: I1203 00:20:58.042197 4953 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"elastic-operator-dockercfg-4d2wl" Dec 03 00:20:58 crc kubenswrapper[4953]: I1203 00:20:58.042410 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"openshift-service-ca.crt" Dec 03 00:20:58 crc kubenswrapper[4953]: I1203 00:20:58.067466 4953 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/elastic-operator-866854d65-qhl5q"] Dec 03 00:20:58 crc kubenswrapper[4953]: I1203 00:20:58.072901 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/3eb60698-d3b7-4fad-9b46-c45ad8b132a7-apiservice-cert\") pod \"elastic-operator-866854d65-qhl5q\" (UID: \"3eb60698-d3b7-4fad-9b46-c45ad8b132a7\") " pod="service-telemetry/elastic-operator-866854d65-qhl5q" Dec 03 00:20:58 crc kubenswrapper[4953]: I1203 00:20:58.072991 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tgqvh\" (UniqueName: \"kubernetes.io/projected/3eb60698-d3b7-4fad-9b46-c45ad8b132a7-kube-api-access-tgqvh\") pod \"elastic-operator-866854d65-qhl5q\" (UID: \"3eb60698-d3b7-4fad-9b46-c45ad8b132a7\") " pod="service-telemetry/elastic-operator-866854d65-qhl5q" Dec 03 00:20:58 crc kubenswrapper[4953]: I1203 00:20:58.073377 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/3eb60698-d3b7-4fad-9b46-c45ad8b132a7-webhook-cert\") pod \"elastic-operator-866854d65-qhl5q\" (UID: \"3eb60698-d3b7-4fad-9b46-c45ad8b132a7\") " pod="service-telemetry/elastic-operator-866854d65-qhl5q" Dec 03 00:20:58 crc kubenswrapper[4953]: I1203 00:20:58.173465 4953 generic.go:334] "Generic (PLEG): container finished" podID="37d22f65-99db-40b6-85ce-a22f2b448708" containerID="3090ad29f26f813c98f5d2f7f67d2ff61214d101a507315b4456f25138670998" exitCode=0 Dec 03 00:20:58 crc kubenswrapper[4953]: I1203 00:20:58.173542 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-gglz9" event={"ID":"37d22f65-99db-40b6-85ce-a22f2b448708","Type":"ContainerDied","Data":"3090ad29f26f813c98f5d2f7f67d2ff61214d101a507315b4456f25138670998"} Dec 03 00:20:58 crc kubenswrapper[4953]: I1203 00:20:58.179020 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/3eb60698-d3b7-4fad-9b46-c45ad8b132a7-webhook-cert\") pod \"elastic-operator-866854d65-qhl5q\" (UID: \"3eb60698-d3b7-4fad-9b46-c45ad8b132a7\") " pod="service-telemetry/elastic-operator-866854d65-qhl5q" Dec 03 00:20:58 crc kubenswrapper[4953]: I1203 00:20:58.179117 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/3eb60698-d3b7-4fad-9b46-c45ad8b132a7-apiservice-cert\") pod \"elastic-operator-866854d65-qhl5q\" (UID: \"3eb60698-d3b7-4fad-9b46-c45ad8b132a7\") " pod="service-telemetry/elastic-operator-866854d65-qhl5q" Dec 03 00:20:58 crc kubenswrapper[4953]: I1203 00:20:58.179161 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tgqvh\" (UniqueName: \"kubernetes.io/projected/3eb60698-d3b7-4fad-9b46-c45ad8b132a7-kube-api-access-tgqvh\") pod \"elastic-operator-866854d65-qhl5q\" (UID: \"3eb60698-d3b7-4fad-9b46-c45ad8b132a7\") " pod="service-telemetry/elastic-operator-866854d65-qhl5q" Dec 03 00:20:58 crc kubenswrapper[4953]: I1203 00:20:58.192484 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/3eb60698-d3b7-4fad-9b46-c45ad8b132a7-apiservice-cert\") pod \"elastic-operator-866854d65-qhl5q\" (UID: \"3eb60698-d3b7-4fad-9b46-c45ad8b132a7\") " pod="service-telemetry/elastic-operator-866854d65-qhl5q" Dec 03 00:20:58 crc kubenswrapper[4953]: I1203 00:20:58.192870 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/3eb60698-d3b7-4fad-9b46-c45ad8b132a7-webhook-cert\") pod \"elastic-operator-866854d65-qhl5q\" (UID: \"3eb60698-d3b7-4fad-9b46-c45ad8b132a7\") " pod="service-telemetry/elastic-operator-866854d65-qhl5q" Dec 03 00:20:58 crc kubenswrapper[4953]: I1203 00:20:58.215940 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tgqvh\" (UniqueName: \"kubernetes.io/projected/3eb60698-d3b7-4fad-9b46-c45ad8b132a7-kube-api-access-tgqvh\") pod \"elastic-operator-866854d65-qhl5q\" (UID: \"3eb60698-d3b7-4fad-9b46-c45ad8b132a7\") " pod="service-telemetry/elastic-operator-866854d65-qhl5q" Dec 03 00:20:58 crc kubenswrapper[4953]: I1203 00:20:58.362123 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/elastic-operator-866854d65-qhl5q" Dec 03 00:20:59 crc kubenswrapper[4953]: I1203 00:20:59.751245 4953 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["service-telemetry/interconnect-operator-5bb49f789d-q28d5"] Dec 03 00:20:59 crc kubenswrapper[4953]: I1203 00:20:59.752322 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/interconnect-operator-5bb49f789d-q28d5" Dec 03 00:20:59 crc kubenswrapper[4953]: I1203 00:20:59.756918 4953 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/interconnect-operator-5bb49f789d-q28d5"] Dec 03 00:20:59 crc kubenswrapper[4953]: I1203 00:20:59.757375 4953 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"interconnect-operator-dockercfg-nchh5" Dec 03 00:20:59 crc kubenswrapper[4953]: I1203 00:20:59.893190 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4b9t5\" (UniqueName: \"kubernetes.io/projected/4ae35630-6002-4c86-84ed-135aeda64570-kube-api-access-4b9t5\") pod \"interconnect-operator-5bb49f789d-q28d5\" (UID: \"4ae35630-6002-4c86-84ed-135aeda64570\") " pod="service-telemetry/interconnect-operator-5bb49f789d-q28d5" Dec 03 00:21:00 crc kubenswrapper[4953]: I1203 00:21:00.009088 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4b9t5\" (UniqueName: \"kubernetes.io/projected/4ae35630-6002-4c86-84ed-135aeda64570-kube-api-access-4b9t5\") pod \"interconnect-operator-5bb49f789d-q28d5\" (UID: \"4ae35630-6002-4c86-84ed-135aeda64570\") " pod="service-telemetry/interconnect-operator-5bb49f789d-q28d5" Dec 03 00:21:00 crc kubenswrapper[4953]: I1203 00:21:00.043215 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4b9t5\" (UniqueName: \"kubernetes.io/projected/4ae35630-6002-4c86-84ed-135aeda64570-kube-api-access-4b9t5\") pod \"interconnect-operator-5bb49f789d-q28d5\" (UID: \"4ae35630-6002-4c86-84ed-135aeda64570\") " pod="service-telemetry/interconnect-operator-5bb49f789d-q28d5" Dec 03 00:21:00 crc kubenswrapper[4953]: I1203 00:21:00.137030 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/interconnect-operator-5bb49f789d-q28d5" Dec 03 00:21:03 crc kubenswrapper[4953]: E1203 00:21:03.083236 4953 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 3090ad29f26f813c98f5d2f7f67d2ff61214d101a507315b4456f25138670998 is running failed: container process not found" containerID="3090ad29f26f813c98f5d2f7f67d2ff61214d101a507315b4456f25138670998" cmd=["grpc_health_probe","-addr=:50051"] Dec 03 00:21:03 crc kubenswrapper[4953]: E1203 00:21:03.084938 4953 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 3090ad29f26f813c98f5d2f7f67d2ff61214d101a507315b4456f25138670998 is running failed: container process not found" containerID="3090ad29f26f813c98f5d2f7f67d2ff61214d101a507315b4456f25138670998" cmd=["grpc_health_probe","-addr=:50051"] Dec 03 00:21:03 crc kubenswrapper[4953]: E1203 00:21:03.085875 4953 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 3090ad29f26f813c98f5d2f7f67d2ff61214d101a507315b4456f25138670998 is running failed: container process not found" containerID="3090ad29f26f813c98f5d2f7f67d2ff61214d101a507315b4456f25138670998" cmd=["grpc_health_probe","-addr=:50051"] Dec 03 00:21:03 crc kubenswrapper[4953]: E1203 00:21:03.085919 4953 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 3090ad29f26f813c98f5d2f7f67d2ff61214d101a507315b4456f25138670998 is running failed: container process not found" probeType="Readiness" pod="openshift-marketplace/certified-operators-gglz9" podUID="37d22f65-99db-40b6-85ce-a22f2b448708" containerName="registry-server" Dec 03 00:21:05 crc kubenswrapper[4953]: E1203 00:21:05.931921 4953 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="registry.redhat.io/cert-manager/cert-manager-operator-bundle@sha256:acaaea813059d4ac5b2618395bd9113f72ada0a33aaaba91aa94f000e77df407" Dec 03 00:21:05 crc kubenswrapper[4953]: E1203 00:21:05.932709 4953 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:pull,Image:registry.redhat.io/cert-manager/cert-manager-operator-bundle@sha256:acaaea813059d4ac5b2618395bd9113f72ada0a33aaaba91aa94f000e77df407,Command:[/util/cpb /bundle],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{52428800 0} {} 50Mi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:bundle,ReadOnly:false,MountPath:/bundle,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:util,ReadOnly:false,MountPath:/util,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-58gm2,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod 1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931ahqk5h_openshift-marketplace(8bc0877b-282d-457a-a46a-c3202a31e2fb): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 03 00:21:05 crc kubenswrapper[4953]: E1203 00:21:05.934321 4953 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"pull\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931ahqk5h" podUID="8bc0877b-282d-457a-a46a-c3202a31e2fb" Dec 03 00:21:06 crc kubenswrapper[4953]: E1203 00:21:06.446783 4953 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"pull\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/cert-manager/cert-manager-operator-bundle@sha256:acaaea813059d4ac5b2618395bd9113f72ada0a33aaaba91aa94f000e77df407\\\"\"" pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931ahqk5h" podUID="8bc0877b-282d-457a-a46a-c3202a31e2fb" Dec 03 00:21:13 crc kubenswrapper[4953]: E1203 00:21:13.081893 4953 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 3090ad29f26f813c98f5d2f7f67d2ff61214d101a507315b4456f25138670998 is running failed: container process not found" containerID="3090ad29f26f813c98f5d2f7f67d2ff61214d101a507315b4456f25138670998" cmd=["grpc_health_probe","-addr=:50051"] Dec 03 00:21:13 crc kubenswrapper[4953]: E1203 00:21:13.083512 4953 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 3090ad29f26f813c98f5d2f7f67d2ff61214d101a507315b4456f25138670998 is running failed: container process not found" containerID="3090ad29f26f813c98f5d2f7f67d2ff61214d101a507315b4456f25138670998" cmd=["grpc_health_probe","-addr=:50051"] Dec 03 00:21:13 crc kubenswrapper[4953]: E1203 00:21:13.089666 4953 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 3090ad29f26f813c98f5d2f7f67d2ff61214d101a507315b4456f25138670998 is running failed: container process not found" containerID="3090ad29f26f813c98f5d2f7f67d2ff61214d101a507315b4456f25138670998" cmd=["grpc_health_probe","-addr=:50051"] Dec 03 00:21:13 crc kubenswrapper[4953]: E1203 00:21:13.089785 4953 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 3090ad29f26f813c98f5d2f7f67d2ff61214d101a507315b4456f25138670998 is running failed: container process not found" probeType="Readiness" pod="openshift-marketplace/certified-operators-gglz9" podUID="37d22f65-99db-40b6-85ce-a22f2b448708" containerName="registry-server" Dec 03 00:21:13 crc kubenswrapper[4953]: I1203 00:21:13.793311 4953 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-gglz9" Dec 03 00:21:13 crc kubenswrapper[4953]: I1203 00:21:13.916998 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/37d22f65-99db-40b6-85ce-a22f2b448708-catalog-content\") pod \"37d22f65-99db-40b6-85ce-a22f2b448708\" (UID: \"37d22f65-99db-40b6-85ce-a22f2b448708\") " Dec 03 00:21:13 crc kubenswrapper[4953]: I1203 00:21:13.917178 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/37d22f65-99db-40b6-85ce-a22f2b448708-utilities\") pod \"37d22f65-99db-40b6-85ce-a22f2b448708\" (UID: \"37d22f65-99db-40b6-85ce-a22f2b448708\") " Dec 03 00:21:13 crc kubenswrapper[4953]: I1203 00:21:13.917227 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ljp92\" (UniqueName: \"kubernetes.io/projected/37d22f65-99db-40b6-85ce-a22f2b448708-kube-api-access-ljp92\") pod \"37d22f65-99db-40b6-85ce-a22f2b448708\" (UID: \"37d22f65-99db-40b6-85ce-a22f2b448708\") " Dec 03 00:21:13 crc kubenswrapper[4953]: I1203 00:21:13.918732 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/37d22f65-99db-40b6-85ce-a22f2b448708-utilities" (OuterVolumeSpecName: "utilities") pod "37d22f65-99db-40b6-85ce-a22f2b448708" (UID: "37d22f65-99db-40b6-85ce-a22f2b448708"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 00:21:13 crc kubenswrapper[4953]: I1203 00:21:13.929701 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/37d22f65-99db-40b6-85ce-a22f2b448708-kube-api-access-ljp92" (OuterVolumeSpecName: "kube-api-access-ljp92") pod "37d22f65-99db-40b6-85ce-a22f2b448708" (UID: "37d22f65-99db-40b6-85ce-a22f2b448708"). InnerVolumeSpecName "kube-api-access-ljp92". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:21:13 crc kubenswrapper[4953]: I1203 00:21:13.982341 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/37d22f65-99db-40b6-85ce-a22f2b448708-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "37d22f65-99db-40b6-85ce-a22f2b448708" (UID: "37d22f65-99db-40b6-85ce-a22f2b448708"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 00:21:14 crc kubenswrapper[4953]: I1203 00:21:14.019232 4953 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/37d22f65-99db-40b6-85ce-a22f2b448708-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 00:21:14 crc kubenswrapper[4953]: I1203 00:21:14.019300 4953 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ljp92\" (UniqueName: \"kubernetes.io/projected/37d22f65-99db-40b6-85ce-a22f2b448708-kube-api-access-ljp92\") on node \"crc\" DevicePath \"\"" Dec 03 00:21:14 crc kubenswrapper[4953]: I1203 00:21:14.019315 4953 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/37d22f65-99db-40b6-85ce-a22f2b448708-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 00:21:14 crc kubenswrapper[4953]: I1203 00:21:14.555231 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-gglz9" event={"ID":"37d22f65-99db-40b6-85ce-a22f2b448708","Type":"ContainerDied","Data":"53a10dec83122a9e74fac5ca93d650561cd19cab7f08d9cb75f8c200d59a2002"} Dec 03 00:21:14 crc kubenswrapper[4953]: I1203 00:21:14.555285 4953 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-gglz9" Dec 03 00:21:14 crc kubenswrapper[4953]: I1203 00:21:14.555319 4953 scope.go:117] "RemoveContainer" containerID="3090ad29f26f813c98f5d2f7f67d2ff61214d101a507315b4456f25138670998" Dec 03 00:21:14 crc kubenswrapper[4953]: I1203 00:21:14.587393 4953 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-gglz9"] Dec 03 00:21:14 crc kubenswrapper[4953]: I1203 00:21:14.592575 4953 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-gglz9"] Dec 03 00:21:14 crc kubenswrapper[4953]: I1203 00:21:14.680730 4953 scope.go:117] "RemoveContainer" containerID="0a4fd3ee636c0a87cb2d03c8b75bf58b63c94bf55396b1c07ca8d25ae0ee92b1" Dec 03 00:21:14 crc kubenswrapper[4953]: E1203 00:21:14.684576 4953 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/cluster-observability-operator/perses-rhel9-operator@sha256:9aec4c328ec43e40481e06ca5808deead74b75c0aacb90e9e72966c3fa14f385" Dec 03 00:21:14 crc kubenswrapper[4953]: E1203 00:21:14.684907 4953 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:perses-operator,Image:registry.redhat.io/cluster-observability-operator/perses-rhel9-operator@sha256:9aec4c328ec43e40481e06ca5808deead74b75c0aacb90e9e72966c3fa14f385,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:OPERATOR_CONDITION_NAME,Value:cluster-observability-operator.v1.3.0,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{100 -3} {} 100m DecimalSI},memory: {{134217728 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:openshift-service-ca,ReadOnly:true,MountPath:/ca,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-zckhl,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000350000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod perses-operator-5446b9c989-hqmj8_openshift-operators(61476fab-b1b9-424c-ab5d-853f83d6fdfd): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 03 00:21:14 crc kubenswrapper[4953]: E1203 00:21:14.686171 4953 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"perses-operator\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-operators/perses-operator-5446b9c989-hqmj8" podUID="61476fab-b1b9-424c-ab5d-853f83d6fdfd" Dec 03 00:21:15 crc kubenswrapper[4953]: E1203 00:21:15.281625 4953 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/cluster-observability-operator/obo-prometheus-operator-admission-webhook-rhel9@sha256:43d33f0125e6b990f4a972ac4e952a065d7e72dc1690c6c836963b7341734aec" Dec 03 00:21:15 crc kubenswrapper[4953]: E1203 00:21:15.281929 4953 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:prometheus-operator-admission-webhook,Image:registry.redhat.io/cluster-observability-operator/obo-prometheus-operator-admission-webhook-rhel9@sha256:43d33f0125e6b990f4a972ac4e952a065d7e72dc1690c6c836963b7341734aec,Command:[],Args:[--web.enable-tls=true --web.cert-file=/tmp/k8s-webhook-server/serving-certs/tls.crt --web.key-file=/tmp/k8s-webhook-server/serving-certs/tls.key],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:OPERATOR_CONDITION_NAME,Value:cluster-observability-operator.v1.3.0,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{200 -3} {} 200m DecimalSI},memory: {{209715200 0} {} BinarySI},},Requests:ResourceList{cpu: {{50 -3} {} 50m DecimalSI},memory: {{52428800 0} {} 50Mi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:apiservice-cert,ReadOnly:false,MountPath:/apiserver.local.config/certificates,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:webhook-cert,ReadOnly:false,MountPath:/tmp/k8s-webhook-server/serving-certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:*true,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod obo-prometheus-operator-admission-webhook-68bbd5c68d-2lg5v_openshift-operators(0ce8b8d2-06e2-4df3-9a3f-d2022263fd97): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 03 00:21:15 crc kubenswrapper[4953]: E1203 00:21:15.283105 4953 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"prometheus-operator-admission-webhook\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-operators/obo-prometheus-operator-admission-webhook-68bbd5c68d-2lg5v" podUID="0ce8b8d2-06e2-4df3-9a3f-d2022263fd97" Dec 03 00:21:15 crc kubenswrapper[4953]: I1203 00:21:15.360555 4953 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="37d22f65-99db-40b6-85ce-a22f2b448708" path="/var/lib/kubelet/pods/37d22f65-99db-40b6-85ce-a22f2b448708/volumes" Dec 03 00:21:15 crc kubenswrapper[4953]: E1203 00:21:15.629393 4953 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"prometheus-operator-admission-webhook\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/cluster-observability-operator/obo-prometheus-operator-admission-webhook-rhel9@sha256:43d33f0125e6b990f4a972ac4e952a065d7e72dc1690c6c836963b7341734aec\\\"\"" pod="openshift-operators/obo-prometheus-operator-admission-webhook-68bbd5c68d-2lg5v" podUID="0ce8b8d2-06e2-4df3-9a3f-d2022263fd97" Dec 03 00:21:15 crc kubenswrapper[4953]: E1203 00:21:15.630055 4953 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"perses-operator\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/cluster-observability-operator/perses-rhel9-operator@sha256:9aec4c328ec43e40481e06ca5808deead74b75c0aacb90e9e72966c3fa14f385\\\"\"" pod="openshift-operators/perses-operator-5446b9c989-hqmj8" podUID="61476fab-b1b9-424c-ab5d-853f83d6fdfd" Dec 03 00:21:16 crc kubenswrapper[4953]: E1203 00:21:16.478204 4953 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/cluster-observability-operator/obo-prometheus-rhel9-operator@sha256:203cf5b9dc1460f09e75f58d8b5cf7df5e57c18c8c6a41c14b5e8977d83263f3" Dec 03 00:21:16 crc kubenswrapper[4953]: E1203 00:21:16.479174 4953 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:prometheus-operator,Image:registry.redhat.io/cluster-observability-operator/obo-prometheus-rhel9-operator@sha256:203cf5b9dc1460f09e75f58d8b5cf7df5e57c18c8c6a41c14b5e8977d83263f3,Command:[],Args:[--prometheus-config-reloader=$(RELATED_IMAGE_PROMETHEUS_CONFIG_RELOADER) --prometheus-instance-selector=app.kubernetes.io/managed-by=observability-operator --alertmanager-instance-selector=app.kubernetes.io/managed-by=observability-operator --thanos-ruler-instance-selector=app.kubernetes.io/managed-by=observability-operator],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:http,HostPort:0,ContainerPort:8080,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:GOGC,Value:30,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_PROMETHEUS_CONFIG_RELOADER,Value:registry.redhat.io/cluster-observability-operator/obo-prometheus-operator-prometheus-config-reloader-rhel9@sha256:1133c973c7472c665f910a722e19c8e2e27accb34b90fab67f14548627ce9c62,ValueFrom:nil,},EnvVar{Name:OPERATOR_CONDITION_NAME,Value:cluster-observability-operator.v1.3.0,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{100 -3} {} 100m DecimalSI},memory: {{524288000 0} {} 500Mi BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{157286400 0} {} 150Mi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-64956,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:*true,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod obo-prometheus-operator-668cf9dfbb-kk8d4_openshift-operators(f5d63f17-0a24-453c-a528-e61266f2b091): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 03 00:21:16 crc kubenswrapper[4953]: E1203 00:21:16.480404 4953 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"prometheus-operator\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-kk8d4" podUID="f5d63f17-0a24-453c-a528-e61266f2b091" Dec 03 00:21:16 crc kubenswrapper[4953]: E1203 00:21:16.639429 4953 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"prometheus-operator\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/cluster-observability-operator/obo-prometheus-rhel9-operator@sha256:203cf5b9dc1460f09e75f58d8b5cf7df5e57c18c8c6a41c14b5e8977d83263f3\\\"\"" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-kk8d4" podUID="f5d63f17-0a24-453c-a528-e61266f2b091" Dec 03 00:21:18 crc kubenswrapper[4953]: I1203 00:21:18.921461 4953 scope.go:117] "RemoveContainer" containerID="15e4958e6f63fcd52549111ef31ed1df48c171486f7a5f1de8354a51d90ea9af" Dec 03 00:21:18 crc kubenswrapper[4953]: E1203 00:21:18.995030 4953 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/cluster-observability-operator/cluster-observability-rhel9-operator@sha256:ce7d2904f7b238aa37dfe74a0b76bf73629e7a14fa52bf54b0ecf030ca36f1bb" Dec 03 00:21:18 crc kubenswrapper[4953]: E1203 00:21:18.995516 4953 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:operator,Image:registry.redhat.io/cluster-observability-operator/cluster-observability-rhel9-operator@sha256:ce7d2904f7b238aa37dfe74a0b76bf73629e7a14fa52bf54b0ecf030ca36f1bb,Command:[],Args:[--namespace=$(NAMESPACE) --images=perses=$(RELATED_IMAGE_PERSES) --images=alertmanager=$(RELATED_IMAGE_ALERTMANAGER) --images=prometheus=$(RELATED_IMAGE_PROMETHEUS) --images=thanos=$(RELATED_IMAGE_THANOS) --images=ui-dashboards=$(RELATED_IMAGE_CONSOLE_DASHBOARDS_PLUGIN) --images=ui-distributed-tracing=$(RELATED_IMAGE_CONSOLE_DISTRIBUTED_TRACING_PLUGIN) --images=ui-distributed-tracing-pf5=$(RELATED_IMAGE_CONSOLE_DISTRIBUTED_TRACING_PLUGIN_PF5) --images=ui-distributed-tracing-pf4=$(RELATED_IMAGE_CONSOLE_DISTRIBUTED_TRACING_PLUGIN_PF4) --images=ui-logging=$(RELATED_IMAGE_CONSOLE_LOGGING_PLUGIN) --images=ui-logging-pf4=$(RELATED_IMAGE_CONSOLE_LOGGING_PLUGIN_PF4) --images=ui-troubleshooting-panel=$(RELATED_IMAGE_CONSOLE_TROUBLESHOOTING_PANEL_PLUGIN) --images=ui-monitoring=$(RELATED_IMAGE_CONSOLE_MONITORING_PLUGIN) --images=ui-monitoring-pf5=$(RELATED_IMAGE_CONSOLE_MONITORING_PLUGIN_PF5) --images=korrel8r=$(RELATED_IMAGE_KORREL8R) --images=health-analyzer=$(RELATED_IMAGE_CLUSTER_HEALTH_ANALYZER) --openshift.enabled=true],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:NAMESPACE,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:RELATED_IMAGE_ALERTMANAGER,Value:registry.redhat.io/cluster-observability-operator/alertmanager-rhel9@sha256:e718854a7d6ca8accf0fa72db0eb902e46c44d747ad51dc3f06bba0cefaa3c01,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_PROMETHEUS,Value:registry.redhat.io/cluster-observability-operator/prometheus-rhel9@sha256:17ea20be390a94ab39f5cdd7f0cbc2498046eebcf77fe3dec9aa288d5c2cf46b,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_THANOS,Value:registry.redhat.io/cluster-observability-operator/thanos-rhel9@sha256:d972f4faa5e9c121402d23ed85002f26af48ec36b1b71a7489d677b3913d08b4,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_PERSES,Value:registry.redhat.io/cluster-observability-operator/perses-rhel9@sha256:91531137fc1dcd740e277e0f65e120a0176a16f788c14c27925b61aa0b792ade,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CONSOLE_DASHBOARDS_PLUGIN,Value:registry.redhat.io/cluster-observability-operator/dashboards-console-plugin-rhel9@sha256:a69da8bbca8a28dd2925f864d51cc31cf761b10532c553095ba40b242ef701cb,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CONSOLE_DISTRIBUTED_TRACING_PLUGIN,Value:registry.redhat.io/cluster-observability-operator/distributed-tracing-console-plugin-rhel9@sha256:897e1bfad1187062725b54d87107bd0155972257a50d8335dd29e1999b828a4f,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CONSOLE_DISTRIBUTED_TRACING_PLUGIN_PF5,Value:registry.redhat.io/cluster-observability-operator/distributed-tracing-console-plugin-pf5-rhel9@sha256:95fe5b5746ca8c07ac9217ce2d8ac8e6afad17af210f9d8e0074df1310b209a8,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CONSOLE_DISTRIBUTED_TRACING_PLUGIN_PF4,Value:registry.redhat.io/cluster-observability-operator/distributed-tracing-console-plugin-pf4-rhel9@sha256:e9d9a89e4d8126a62b1852055482258ee528cac6398dd5d43ebad75ace0f33c9,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CONSOLE_LOGGING_PLUGIN,Value:registry.redhat.io/cluster-observability-operator/logging-console-plugin-rhel9@sha256:ec684a0645ceb917b019af7ddba68c3533416e356ab0d0320a30e75ca7ebb31b,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CONSOLE_LOGGING_PLUGIN_PF4,Value:registry.redhat.io/cluster-observability-operator/logging-console-plugin-pf4-rhel9@sha256:3b9693fcde9b3a9494fb04735b1f7cfd0426f10be820fdc3f024175c0d3df1c9,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CONSOLE_TROUBLESHOOTING_PANEL_PLUGIN,Value:registry.redhat.io/cluster-observability-operator/troubleshooting-panel-console-plugin-rhel9@sha256:580606f194180accc8abba099e17a26dca7522ec6d233fa2fdd40312771703e3,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CONSOLE_MONITORING_PLUGIN,Value:registry.redhat.io/cluster-observability-operator/monitoring-console-plugin-rhel9@sha256:e03777be39e71701935059cd877603874a13ac94daa73219d4e5e545599d78a9,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CONSOLE_MONITORING_PLUGIN_PF5,Value:registry.redhat.io/cluster-observability-operator/monitoring-console-plugin-pf5-rhel9@sha256:aa47256193cfd2877853878e1ae97d2ab8b8e5deae62b387cbfad02b284d379c,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_KORREL8R,Value:registry.redhat.io/cluster-observability-operator/korrel8r-rhel9@sha256:c595ff56b2cb85514bf4784db6ddb82e4e657e3e708a7fb695fc4997379a94d4,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CLUSTER_HEALTH_ANALYZER,Value:registry.redhat.io/cluster-observability-operator/cluster-health-analyzer-rhel9@sha256:45a4ec2a519bcec99e886aa91596d5356a2414a2bd103baaef9fa7838c672eb2,ValueFrom:nil,},EnvVar{Name:OPERATOR_CONDITION_NAME,Value:cluster-observability-operator.v1.3.0,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{400 -3} {} 400m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{100 -3} {} 100m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:observability-operator-tls,ReadOnly:true,MountPath:/etc/tls/private,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-9dn9l,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000350000,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod observability-operator-d8bb48f5d-64xzt_openshift-operators(1df642b8-586a-4bc5-8d9f-9f2cff7c6b80): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 03 00:21:18 crc kubenswrapper[4953]: E1203 00:21:18.996054 4953 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/cluster-observability-operator/obo-prometheus-operator-admission-webhook-rhel9@sha256:43d33f0125e6b990f4a972ac4e952a065d7e72dc1690c6c836963b7341734aec" Dec 03 00:21:18 crc kubenswrapper[4953]: E1203 00:21:18.996169 4953 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:prometheus-operator-admission-webhook,Image:registry.redhat.io/cluster-observability-operator/obo-prometheus-operator-admission-webhook-rhel9@sha256:43d33f0125e6b990f4a972ac4e952a065d7e72dc1690c6c836963b7341734aec,Command:[],Args:[--web.enable-tls=true --web.cert-file=/tmp/k8s-webhook-server/serving-certs/tls.crt --web.key-file=/tmp/k8s-webhook-server/serving-certs/tls.key],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:OPERATOR_CONDITION_NAME,Value:cluster-observability-operator.v1.3.0,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{200 -3} {} 200m DecimalSI},memory: {{209715200 0} {} BinarySI},},Requests:ResourceList{cpu: {{50 -3} {} 50m DecimalSI},memory: {{52428800 0} {} 50Mi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:apiservice-cert,ReadOnly:false,MountPath:/apiserver.local.config/certificates,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:webhook-cert,ReadOnly:false,MountPath:/tmp/k8s-webhook-server/serving-certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:*true,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod obo-prometheus-operator-admission-webhook-68bbd5c68d-pmttp_openshift-operators(fc13989b-f854-49fd-8901-3b206167aafc): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 03 00:21:18 crc kubenswrapper[4953]: E1203 00:21:18.997328 4953 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"prometheus-operator-admission-webhook\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-operators/obo-prometheus-operator-admission-webhook-68bbd5c68d-pmttp" podUID="fc13989b-f854-49fd-8901-3b206167aafc" Dec 03 00:21:18 crc kubenswrapper[4953]: E1203 00:21:18.997392 4953 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-operators/observability-operator-d8bb48f5d-64xzt" podUID="1df642b8-586a-4bc5-8d9f-9f2cff7c6b80" Dec 03 00:21:19 crc kubenswrapper[4953]: I1203 00:21:19.272912 4953 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/interconnect-operator-5bb49f789d-q28d5"] Dec 03 00:21:19 crc kubenswrapper[4953]: I1203 00:21:19.319597 4953 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/elastic-operator-866854d65-qhl5q"] Dec 03 00:21:19 crc kubenswrapper[4953]: W1203 00:21:19.321859 4953 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3eb60698_d3b7_4fad_9b46_c45ad8b132a7.slice/crio-94a19656b2324942d4e1b56244227eca18f243aff7aedf32e488373febef54bb WatchSource:0}: Error finding container 94a19656b2324942d4e1b56244227eca18f243aff7aedf32e488373febef54bb: Status 404 returned error can't find the container with id 94a19656b2324942d4e1b56244227eca18f243aff7aedf32e488373febef54bb Dec 03 00:21:19 crc kubenswrapper[4953]: I1203 00:21:19.655418 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/interconnect-operator-5bb49f789d-q28d5" event={"ID":"4ae35630-6002-4c86-84ed-135aeda64570","Type":"ContainerStarted","Data":"f625d2d42c488c3b14d17d3b8dd8052f4aeebdd9f1a7cd34caca72c484c55cb8"} Dec 03 00:21:19 crc kubenswrapper[4953]: I1203 00:21:19.656882 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/elastic-operator-866854d65-qhl5q" event={"ID":"3eb60698-d3b7-4fad-9b46-c45ad8b132a7","Type":"ContainerStarted","Data":"94a19656b2324942d4e1b56244227eca18f243aff7aedf32e488373febef54bb"} Dec 03 00:21:19 crc kubenswrapper[4953]: E1203 00:21:19.660124 4953 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"prometheus-operator-admission-webhook\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/cluster-observability-operator/obo-prometheus-operator-admission-webhook-rhel9@sha256:43d33f0125e6b990f4a972ac4e952a065d7e72dc1690c6c836963b7341734aec\\\"\"" pod="openshift-operators/obo-prometheus-operator-admission-webhook-68bbd5c68d-pmttp" podUID="fc13989b-f854-49fd-8901-3b206167aafc" Dec 03 00:21:19 crc kubenswrapper[4953]: E1203 00:21:19.660149 4953 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/cluster-observability-operator/cluster-observability-rhel9-operator@sha256:ce7d2904f7b238aa37dfe74a0b76bf73629e7a14fa52bf54b0ecf030ca36f1bb\\\"\"" pod="openshift-operators/observability-operator-d8bb48f5d-64xzt" podUID="1df642b8-586a-4bc5-8d9f-9f2cff7c6b80" Dec 03 00:21:21 crc kubenswrapper[4953]: I1203 00:21:21.674051 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931ahqk5h" event={"ID":"8bc0877b-282d-457a-a46a-c3202a31e2fb","Type":"ContainerStarted","Data":"f12cbe7dbb2c86e73b31e7101eccbd77382167f8713996be04b2dc670a1ed79d"} Dec 03 00:21:22 crc kubenswrapper[4953]: I1203 00:21:22.684283 4953 generic.go:334] "Generic (PLEG): container finished" podID="8bc0877b-282d-457a-a46a-c3202a31e2fb" containerID="f12cbe7dbb2c86e73b31e7101eccbd77382167f8713996be04b2dc670a1ed79d" exitCode=0 Dec 03 00:21:22 crc kubenswrapper[4953]: I1203 00:21:22.684355 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931ahqk5h" event={"ID":"8bc0877b-282d-457a-a46a-c3202a31e2fb","Type":"ContainerDied","Data":"f12cbe7dbb2c86e73b31e7101eccbd77382167f8713996be04b2dc670a1ed79d"} Dec 03 00:21:32 crc kubenswrapper[4953]: E1203 00:21:32.813459 4953 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/amq7/amq-interconnect-operator@sha256:a8b621237c872ded2a1d1d948fbebd693429e4a1ced1d7922406241a078d3d43" Dec 03 00:21:32 crc kubenswrapper[4953]: E1203 00:21:32.814400 4953 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:interconnect-operator,Image:registry.redhat.io/amq7/amq-interconnect-operator@sha256:a8b621237c872ded2a1d1d948fbebd693429e4a1ced1d7922406241a078d3d43,Command:[qdr-operator],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:metrics,HostPort:0,ContainerPort:60000,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:WATCH_NAMESPACE,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:POD_NAME,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.name,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:OPERATOR_NAME,Value:qdr-operator,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_QDROUTERD_IMAGE,Value:registry.redhat.io/amq7/amq-interconnect@sha256:31d87473fa684178a694f9ee331d3c80f2653f9533cb65c2a325752166a077e9,ValueFrom:nil,},EnvVar{Name:OPERATOR_CONDITION_NAME,Value:amq7-interconnect-operator.v1.10.20,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-4b9t5,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000670000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod interconnect-operator-5bb49f789d-q28d5_service-telemetry(4ae35630-6002-4c86-84ed-135aeda64570): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 03 00:21:32 crc kubenswrapper[4953]: E1203 00:21:32.815655 4953 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"interconnect-operator\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="service-telemetry/interconnect-operator-5bb49f789d-q28d5" podUID="4ae35630-6002-4c86-84ed-135aeda64570" Dec 03 00:21:33 crc kubenswrapper[4953]: I1203 00:21:33.765781 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/observability-operator-d8bb48f5d-64xzt" event={"ID":"1df642b8-586a-4bc5-8d9f-9f2cff7c6b80","Type":"ContainerStarted","Data":"a73e34c31bccb0d0929a6b3a6662f74796e136519cb5b6410763b6212a5572a0"} Dec 03 00:21:33 crc kubenswrapper[4953]: I1203 00:21:33.767645 4953 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operators/observability-operator-d8bb48f5d-64xzt" Dec 03 00:21:33 crc kubenswrapper[4953]: I1203 00:21:33.767986 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-kk8d4" event={"ID":"f5d63f17-0a24-453c-a528-e61266f2b091","Type":"ContainerStarted","Data":"332eb44016cf7c0a97d086d44afd8caf634509f76e88484bc47309b80432cdcd"} Dec 03 00:21:33 crc kubenswrapper[4953]: I1203 00:21:33.771335 4953 patch_prober.go:28] interesting pod/observability-operator-d8bb48f5d-64xzt container/operator namespace/openshift-operators: Readiness probe status=failure output="Get \"http://10.217.0.47:8081/healthz\": dial tcp 10.217.0.47:8081: connect: connection refused" start-of-body= Dec 03 00:21:33 crc kubenswrapper[4953]: I1203 00:21:33.771399 4953 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operators/observability-operator-d8bb48f5d-64xzt" podUID="1df642b8-586a-4bc5-8d9f-9f2cff7c6b80" containerName="operator" probeResult="failure" output="Get \"http://10.217.0.47:8081/healthz\": dial tcp 10.217.0.47:8081: connect: connection refused" Dec 03 00:21:33 crc kubenswrapper[4953]: I1203 00:21:33.774220 4953 generic.go:334] "Generic (PLEG): container finished" podID="8bc0877b-282d-457a-a46a-c3202a31e2fb" containerID="63ec412a856ded1c1a878b98bf19ef6c9c811b12aeddaa4f75933d1646b85de1" exitCode=0 Dec 03 00:21:33 crc kubenswrapper[4953]: I1203 00:21:33.774258 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931ahqk5h" event={"ID":"8bc0877b-282d-457a-a46a-c3202a31e2fb","Type":"ContainerDied","Data":"63ec412a856ded1c1a878b98bf19ef6c9c811b12aeddaa4f75933d1646b85de1"} Dec 03 00:21:33 crc kubenswrapper[4953]: I1203 00:21:33.777288 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/elastic-operator-866854d65-qhl5q" event={"ID":"3eb60698-d3b7-4fad-9b46-c45ad8b132a7","Type":"ContainerStarted","Data":"8920258e414105cb0be62fa5fd6e6538e246df5c48a674d2185a8ce600bd7236"} Dec 03 00:21:33 crc kubenswrapper[4953]: I1203 00:21:33.788793 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/perses-operator-5446b9c989-hqmj8" event={"ID":"61476fab-b1b9-424c-ab5d-853f83d6fdfd","Type":"ContainerStarted","Data":"946288577f0b47bf6ae667723a070879ac31a86d4dab370e3e3cf9dfc4454c59"} Dec 03 00:21:33 crc kubenswrapper[4953]: I1203 00:21:33.789607 4953 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operators/perses-operator-5446b9c989-hqmj8" Dec 03 00:21:33 crc kubenswrapper[4953]: I1203 00:21:33.793866 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-68bbd5c68d-2lg5v" event={"ID":"0ce8b8d2-06e2-4df3-9a3f-d2022263fd97","Type":"ContainerStarted","Data":"79b9e55c860317ca77be2fb8933f7c26737e28b310b11147c0c4a5a5643696a5"} Dec 03 00:21:33 crc kubenswrapper[4953]: E1203 00:21:33.795279 4953 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"interconnect-operator\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/amq7/amq-interconnect-operator@sha256:a8b621237c872ded2a1d1d948fbebd693429e4a1ced1d7922406241a078d3d43\\\"\"" pod="service-telemetry/interconnect-operator-5bb49f789d-q28d5" podUID="4ae35630-6002-4c86-84ed-135aeda64570" Dec 03 00:21:33 crc kubenswrapper[4953]: I1203 00:21:33.802912 4953 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/observability-operator-d8bb48f5d-64xzt" podStartSLOduration=1.952806329 podStartE2EDuration="44.802886558s" podCreationTimestamp="2025-12-03 00:20:49 +0000 UTC" firstStartedPulling="2025-12-03 00:20:50.731063452 +0000 UTC m=+839.715344235" lastFinishedPulling="2025-12-03 00:21:33.581143681 +0000 UTC m=+882.565424464" observedRunningTime="2025-12-03 00:21:33.799107914 +0000 UTC m=+882.783388707" watchObservedRunningTime="2025-12-03 00:21:33.802886558 +0000 UTC m=+882.787167341" Dec 03 00:21:33 crc kubenswrapper[4953]: I1203 00:21:33.824817 4953 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-kk8d4" podStartSLOduration=3.283589847 podStartE2EDuration="45.82478488s" podCreationTimestamp="2025-12-03 00:20:48 +0000 UTC" firstStartedPulling="2025-12-03 00:20:50.390564872 +0000 UTC m=+839.374845655" lastFinishedPulling="2025-12-03 00:21:32.931759905 +0000 UTC m=+881.916040688" observedRunningTime="2025-12-03 00:21:33.82236312 +0000 UTC m=+882.806643913" watchObservedRunningTime="2025-12-03 00:21:33.82478488 +0000 UTC m=+882.809065663" Dec 03 00:21:33 crc kubenswrapper[4953]: I1203 00:21:33.853282 4953 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="service-telemetry/elastic-operator-866854d65-qhl5q" podStartSLOduration=22.24860714 podStartE2EDuration="35.853266036s" podCreationTimestamp="2025-12-03 00:20:58 +0000 UTC" firstStartedPulling="2025-12-03 00:21:19.326152665 +0000 UTC m=+868.310433448" lastFinishedPulling="2025-12-03 00:21:32.930811561 +0000 UTC m=+881.915092344" observedRunningTime="2025-12-03 00:21:33.851673527 +0000 UTC m=+882.835954310" watchObservedRunningTime="2025-12-03 00:21:33.853266036 +0000 UTC m=+882.837546819" Dec 03 00:21:33 crc kubenswrapper[4953]: I1203 00:21:33.929691 4953 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/obo-prometheus-operator-admission-webhook-68bbd5c68d-2lg5v" podStartSLOduration=2.983268141 podStartE2EDuration="44.92965801s" podCreationTimestamp="2025-12-03 00:20:49 +0000 UTC" firstStartedPulling="2025-12-03 00:20:50.9855466 +0000 UTC m=+839.969827383" lastFinishedPulling="2025-12-03 00:21:32.931936469 +0000 UTC m=+881.916217252" observedRunningTime="2025-12-03 00:21:33.925652791 +0000 UTC m=+882.909933574" watchObservedRunningTime="2025-12-03 00:21:33.92965801 +0000 UTC m=+882.913938793" Dec 03 00:21:33 crc kubenswrapper[4953]: I1203 00:21:33.952210 4953 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/perses-operator-5446b9c989-hqmj8" podStartSLOduration=3.01307314 podStartE2EDuration="44.952191669s" podCreationTimestamp="2025-12-03 00:20:49 +0000 UTC" firstStartedPulling="2025-12-03 00:20:50.998770728 +0000 UTC m=+839.983051501" lastFinishedPulling="2025-12-03 00:21:32.937889247 +0000 UTC m=+881.922170030" observedRunningTime="2025-12-03 00:21:33.950801064 +0000 UTC m=+882.935081847" watchObservedRunningTime="2025-12-03 00:21:33.952191669 +0000 UTC m=+882.936472452" Dec 03 00:21:34 crc kubenswrapper[4953]: I1203 00:21:34.527303 4953 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["service-telemetry/elasticsearch-es-default-0"] Dec 03 00:21:34 crc kubenswrapper[4953]: E1203 00:21:34.527563 4953 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="37d22f65-99db-40b6-85ce-a22f2b448708" containerName="extract-content" Dec 03 00:21:34 crc kubenswrapper[4953]: I1203 00:21:34.527580 4953 state_mem.go:107] "Deleted CPUSet assignment" podUID="37d22f65-99db-40b6-85ce-a22f2b448708" containerName="extract-content" Dec 03 00:21:34 crc kubenswrapper[4953]: E1203 00:21:34.527595 4953 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="37d22f65-99db-40b6-85ce-a22f2b448708" containerName="registry-server" Dec 03 00:21:34 crc kubenswrapper[4953]: I1203 00:21:34.527602 4953 state_mem.go:107] "Deleted CPUSet assignment" podUID="37d22f65-99db-40b6-85ce-a22f2b448708" containerName="registry-server" Dec 03 00:21:34 crc kubenswrapper[4953]: E1203 00:21:34.527613 4953 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="37d22f65-99db-40b6-85ce-a22f2b448708" containerName="extract-utilities" Dec 03 00:21:34 crc kubenswrapper[4953]: I1203 00:21:34.527620 4953 state_mem.go:107] "Deleted CPUSet assignment" podUID="37d22f65-99db-40b6-85ce-a22f2b448708" containerName="extract-utilities" Dec 03 00:21:34 crc kubenswrapper[4953]: I1203 00:21:34.527712 4953 memory_manager.go:354] "RemoveStaleState removing state" podUID="37d22f65-99db-40b6-85ce-a22f2b448708" containerName="registry-server" Dec 03 00:21:34 crc kubenswrapper[4953]: I1203 00:21:34.528424 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/elasticsearch-es-default-0" Dec 03 00:21:34 crc kubenswrapper[4953]: I1203 00:21:34.530838 4953 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"elasticsearch-es-http-certs-internal" Dec 03 00:21:34 crc kubenswrapper[4953]: I1203 00:21:34.532416 4953 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"elasticsearch-es-default-es-config" Dec 03 00:21:34 crc kubenswrapper[4953]: I1203 00:21:34.532570 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"elasticsearch-es-unicast-hosts" Dec 03 00:21:34 crc kubenswrapper[4953]: I1203 00:21:34.532613 4953 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"elasticsearch-es-internal-users" Dec 03 00:21:34 crc kubenswrapper[4953]: I1203 00:21:34.532579 4953 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"elasticsearch-es-remote-ca" Dec 03 00:21:34 crc kubenswrapper[4953]: I1203 00:21:34.532797 4953 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"elasticsearch-es-xpack-file-realm" Dec 03 00:21:34 crc kubenswrapper[4953]: I1203 00:21:34.532949 4953 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"elasticsearch-es-default-es-transport-certs" Dec 03 00:21:34 crc kubenswrapper[4953]: I1203 00:21:34.535709 4953 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"default-dockercfg-9xqdk" Dec 03 00:21:34 crc kubenswrapper[4953]: I1203 00:21:34.536780 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"elasticsearch-es-scripts" Dec 03 00:21:34 crc kubenswrapper[4953]: I1203 00:21:34.555624 4953 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/elasticsearch-es-default-0"] Dec 03 00:21:34 crc kubenswrapper[4953]: I1203 00:21:34.631370 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"elastic-internal-elasticsearch-plugins-local\" (UniqueName: \"kubernetes.io/empty-dir/3313dc1f-c9a1-4617-a9d6-cdf983a9d618-elastic-internal-elasticsearch-plugins-local\") pod \"elasticsearch-es-default-0\" (UID: \"3313dc1f-c9a1-4617-a9d6-cdf983a9d618\") " pod="service-telemetry/elasticsearch-es-default-0" Dec 03 00:21:34 crc kubenswrapper[4953]: I1203 00:21:34.631446 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"elastic-internal-elasticsearch-config\" (UniqueName: \"kubernetes.io/secret/3313dc1f-c9a1-4617-a9d6-cdf983a9d618-elastic-internal-elasticsearch-config\") pod \"elasticsearch-es-default-0\" (UID: \"3313dc1f-c9a1-4617-a9d6-cdf983a9d618\") " pod="service-telemetry/elasticsearch-es-default-0" Dec 03 00:21:34 crc kubenswrapper[4953]: I1203 00:21:34.631594 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"elastic-internal-transport-certificates\" (UniqueName: \"kubernetes.io/secret/3313dc1f-c9a1-4617-a9d6-cdf983a9d618-elastic-internal-transport-certificates\") pod \"elasticsearch-es-default-0\" (UID: \"3313dc1f-c9a1-4617-a9d6-cdf983a9d618\") " pod="service-telemetry/elasticsearch-es-default-0" Dec 03 00:21:34 crc kubenswrapper[4953]: I1203 00:21:34.631619 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmp-volume\" (UniqueName: \"kubernetes.io/empty-dir/3313dc1f-c9a1-4617-a9d6-cdf983a9d618-tmp-volume\") pod \"elasticsearch-es-default-0\" (UID: \"3313dc1f-c9a1-4617-a9d6-cdf983a9d618\") " pod="service-telemetry/elasticsearch-es-default-0" Dec 03 00:21:34 crc kubenswrapper[4953]: I1203 00:21:34.631641 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"elastic-internal-http-certificates\" (UniqueName: \"kubernetes.io/secret/3313dc1f-c9a1-4617-a9d6-cdf983a9d618-elastic-internal-http-certificates\") pod \"elasticsearch-es-default-0\" (UID: \"3313dc1f-c9a1-4617-a9d6-cdf983a9d618\") " pod="service-telemetry/elasticsearch-es-default-0" Dec 03 00:21:34 crc kubenswrapper[4953]: I1203 00:21:34.631675 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"elasticsearch-data\" (UniqueName: \"kubernetes.io/empty-dir/3313dc1f-c9a1-4617-a9d6-cdf983a9d618-elasticsearch-data\") pod \"elasticsearch-es-default-0\" (UID: \"3313dc1f-c9a1-4617-a9d6-cdf983a9d618\") " pod="service-telemetry/elasticsearch-es-default-0" Dec 03 00:21:34 crc kubenswrapper[4953]: I1203 00:21:34.631701 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"elastic-internal-elasticsearch-bin-local\" (UniqueName: \"kubernetes.io/empty-dir/3313dc1f-c9a1-4617-a9d6-cdf983a9d618-elastic-internal-elasticsearch-bin-local\") pod \"elasticsearch-es-default-0\" (UID: \"3313dc1f-c9a1-4617-a9d6-cdf983a9d618\") " pod="service-telemetry/elasticsearch-es-default-0" Dec 03 00:21:34 crc kubenswrapper[4953]: I1203 00:21:34.631829 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"elastic-internal-scripts\" (UniqueName: \"kubernetes.io/configmap/3313dc1f-c9a1-4617-a9d6-cdf983a9d618-elastic-internal-scripts\") pod \"elasticsearch-es-default-0\" (UID: \"3313dc1f-c9a1-4617-a9d6-cdf983a9d618\") " pod="service-telemetry/elasticsearch-es-default-0" Dec 03 00:21:34 crc kubenswrapper[4953]: I1203 00:21:34.631939 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"elastic-internal-elasticsearch-config-local\" (UniqueName: \"kubernetes.io/empty-dir/3313dc1f-c9a1-4617-a9d6-cdf983a9d618-elastic-internal-elasticsearch-config-local\") pod \"elasticsearch-es-default-0\" (UID: \"3313dc1f-c9a1-4617-a9d6-cdf983a9d618\") " pod="service-telemetry/elasticsearch-es-default-0" Dec 03 00:21:34 crc kubenswrapper[4953]: I1203 00:21:34.632070 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"elastic-internal-xpack-file-realm\" (UniqueName: \"kubernetes.io/secret/3313dc1f-c9a1-4617-a9d6-cdf983a9d618-elastic-internal-xpack-file-realm\") pod \"elasticsearch-es-default-0\" (UID: \"3313dc1f-c9a1-4617-a9d6-cdf983a9d618\") " pod="service-telemetry/elasticsearch-es-default-0" Dec 03 00:21:34 crc kubenswrapper[4953]: I1203 00:21:34.632508 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"elastic-internal-probe-user\" (UniqueName: \"kubernetes.io/secret/3313dc1f-c9a1-4617-a9d6-cdf983a9d618-elastic-internal-probe-user\") pod \"elasticsearch-es-default-0\" (UID: \"3313dc1f-c9a1-4617-a9d6-cdf983a9d618\") " pod="service-telemetry/elasticsearch-es-default-0" Dec 03 00:21:34 crc kubenswrapper[4953]: I1203 00:21:34.632590 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"elastic-internal-remote-certificate-authorities\" (UniqueName: \"kubernetes.io/secret/3313dc1f-c9a1-4617-a9d6-cdf983a9d618-elastic-internal-remote-certificate-authorities\") pod \"elasticsearch-es-default-0\" (UID: \"3313dc1f-c9a1-4617-a9d6-cdf983a9d618\") " pod="service-telemetry/elasticsearch-es-default-0" Dec 03 00:21:34 crc kubenswrapper[4953]: I1203 00:21:34.632688 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"downward-api\" (UniqueName: \"kubernetes.io/downward-api/3313dc1f-c9a1-4617-a9d6-cdf983a9d618-downward-api\") pod \"elasticsearch-es-default-0\" (UID: \"3313dc1f-c9a1-4617-a9d6-cdf983a9d618\") " pod="service-telemetry/elasticsearch-es-default-0" Dec 03 00:21:34 crc kubenswrapper[4953]: I1203 00:21:34.632713 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"elasticsearch-logs\" (UniqueName: \"kubernetes.io/empty-dir/3313dc1f-c9a1-4617-a9d6-cdf983a9d618-elasticsearch-logs\") pod \"elasticsearch-es-default-0\" (UID: \"3313dc1f-c9a1-4617-a9d6-cdf983a9d618\") " pod="service-telemetry/elasticsearch-es-default-0" Dec 03 00:21:34 crc kubenswrapper[4953]: I1203 00:21:34.632767 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"elastic-internal-unicast-hosts\" (UniqueName: \"kubernetes.io/configmap/3313dc1f-c9a1-4617-a9d6-cdf983a9d618-elastic-internal-unicast-hosts\") pod \"elasticsearch-es-default-0\" (UID: \"3313dc1f-c9a1-4617-a9d6-cdf983a9d618\") " pod="service-telemetry/elasticsearch-es-default-0" Dec 03 00:21:34 crc kubenswrapper[4953]: I1203 00:21:34.734122 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"elastic-internal-transport-certificates\" (UniqueName: \"kubernetes.io/secret/3313dc1f-c9a1-4617-a9d6-cdf983a9d618-elastic-internal-transport-certificates\") pod \"elasticsearch-es-default-0\" (UID: \"3313dc1f-c9a1-4617-a9d6-cdf983a9d618\") " pod="service-telemetry/elasticsearch-es-default-0" Dec 03 00:21:34 crc kubenswrapper[4953]: I1203 00:21:34.734223 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmp-volume\" (UniqueName: \"kubernetes.io/empty-dir/3313dc1f-c9a1-4617-a9d6-cdf983a9d618-tmp-volume\") pod \"elasticsearch-es-default-0\" (UID: \"3313dc1f-c9a1-4617-a9d6-cdf983a9d618\") " pod="service-telemetry/elasticsearch-es-default-0" Dec 03 00:21:34 crc kubenswrapper[4953]: I1203 00:21:34.734262 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"elastic-internal-http-certificates\" (UniqueName: \"kubernetes.io/secret/3313dc1f-c9a1-4617-a9d6-cdf983a9d618-elastic-internal-http-certificates\") pod \"elasticsearch-es-default-0\" (UID: \"3313dc1f-c9a1-4617-a9d6-cdf983a9d618\") " pod="service-telemetry/elasticsearch-es-default-0" Dec 03 00:21:34 crc kubenswrapper[4953]: I1203 00:21:34.734325 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"elasticsearch-data\" (UniqueName: \"kubernetes.io/empty-dir/3313dc1f-c9a1-4617-a9d6-cdf983a9d618-elasticsearch-data\") pod \"elasticsearch-es-default-0\" (UID: \"3313dc1f-c9a1-4617-a9d6-cdf983a9d618\") " pod="service-telemetry/elasticsearch-es-default-0" Dec 03 00:21:34 crc kubenswrapper[4953]: I1203 00:21:34.734369 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"elastic-internal-elasticsearch-bin-local\" (UniqueName: \"kubernetes.io/empty-dir/3313dc1f-c9a1-4617-a9d6-cdf983a9d618-elastic-internal-elasticsearch-bin-local\") pod \"elasticsearch-es-default-0\" (UID: \"3313dc1f-c9a1-4617-a9d6-cdf983a9d618\") " pod="service-telemetry/elasticsearch-es-default-0" Dec 03 00:21:34 crc kubenswrapper[4953]: I1203 00:21:34.734409 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"elastic-internal-scripts\" (UniqueName: \"kubernetes.io/configmap/3313dc1f-c9a1-4617-a9d6-cdf983a9d618-elastic-internal-scripts\") pod \"elasticsearch-es-default-0\" (UID: \"3313dc1f-c9a1-4617-a9d6-cdf983a9d618\") " pod="service-telemetry/elasticsearch-es-default-0" Dec 03 00:21:34 crc kubenswrapper[4953]: I1203 00:21:34.734454 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"elastic-internal-elasticsearch-config-local\" (UniqueName: \"kubernetes.io/empty-dir/3313dc1f-c9a1-4617-a9d6-cdf983a9d618-elastic-internal-elasticsearch-config-local\") pod \"elasticsearch-es-default-0\" (UID: \"3313dc1f-c9a1-4617-a9d6-cdf983a9d618\") " pod="service-telemetry/elasticsearch-es-default-0" Dec 03 00:21:34 crc kubenswrapper[4953]: I1203 00:21:34.734502 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"elastic-internal-xpack-file-realm\" (UniqueName: \"kubernetes.io/secret/3313dc1f-c9a1-4617-a9d6-cdf983a9d618-elastic-internal-xpack-file-realm\") pod \"elasticsearch-es-default-0\" (UID: \"3313dc1f-c9a1-4617-a9d6-cdf983a9d618\") " pod="service-telemetry/elasticsearch-es-default-0" Dec 03 00:21:34 crc kubenswrapper[4953]: I1203 00:21:34.734551 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"elastic-internal-probe-user\" (UniqueName: \"kubernetes.io/secret/3313dc1f-c9a1-4617-a9d6-cdf983a9d618-elastic-internal-probe-user\") pod \"elasticsearch-es-default-0\" (UID: \"3313dc1f-c9a1-4617-a9d6-cdf983a9d618\") " pod="service-telemetry/elasticsearch-es-default-0" Dec 03 00:21:34 crc kubenswrapper[4953]: I1203 00:21:34.734590 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"elastic-internal-remote-certificate-authorities\" (UniqueName: \"kubernetes.io/secret/3313dc1f-c9a1-4617-a9d6-cdf983a9d618-elastic-internal-remote-certificate-authorities\") pod \"elasticsearch-es-default-0\" (UID: \"3313dc1f-c9a1-4617-a9d6-cdf983a9d618\") " pod="service-telemetry/elasticsearch-es-default-0" Dec 03 00:21:34 crc kubenswrapper[4953]: I1203 00:21:34.734632 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"downward-api\" (UniqueName: \"kubernetes.io/downward-api/3313dc1f-c9a1-4617-a9d6-cdf983a9d618-downward-api\") pod \"elasticsearch-es-default-0\" (UID: \"3313dc1f-c9a1-4617-a9d6-cdf983a9d618\") " pod="service-telemetry/elasticsearch-es-default-0" Dec 03 00:21:34 crc kubenswrapper[4953]: I1203 00:21:34.734661 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"elasticsearch-logs\" (UniqueName: \"kubernetes.io/empty-dir/3313dc1f-c9a1-4617-a9d6-cdf983a9d618-elasticsearch-logs\") pod \"elasticsearch-es-default-0\" (UID: \"3313dc1f-c9a1-4617-a9d6-cdf983a9d618\") " pod="service-telemetry/elasticsearch-es-default-0" Dec 03 00:21:34 crc kubenswrapper[4953]: I1203 00:21:34.734701 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"elastic-internal-unicast-hosts\" (UniqueName: \"kubernetes.io/configmap/3313dc1f-c9a1-4617-a9d6-cdf983a9d618-elastic-internal-unicast-hosts\") pod \"elasticsearch-es-default-0\" (UID: \"3313dc1f-c9a1-4617-a9d6-cdf983a9d618\") " pod="service-telemetry/elasticsearch-es-default-0" Dec 03 00:21:34 crc kubenswrapper[4953]: I1203 00:21:34.734758 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"elastic-internal-elasticsearch-plugins-local\" (UniqueName: \"kubernetes.io/empty-dir/3313dc1f-c9a1-4617-a9d6-cdf983a9d618-elastic-internal-elasticsearch-plugins-local\") pod \"elasticsearch-es-default-0\" (UID: \"3313dc1f-c9a1-4617-a9d6-cdf983a9d618\") " pod="service-telemetry/elasticsearch-es-default-0" Dec 03 00:21:34 crc kubenswrapper[4953]: I1203 00:21:34.734810 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"elastic-internal-elasticsearch-config\" (UniqueName: \"kubernetes.io/secret/3313dc1f-c9a1-4617-a9d6-cdf983a9d618-elastic-internal-elasticsearch-config\") pod \"elasticsearch-es-default-0\" (UID: \"3313dc1f-c9a1-4617-a9d6-cdf983a9d618\") " pod="service-telemetry/elasticsearch-es-default-0" Dec 03 00:21:34 crc kubenswrapper[4953]: I1203 00:21:34.734999 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"elasticsearch-data\" (UniqueName: \"kubernetes.io/empty-dir/3313dc1f-c9a1-4617-a9d6-cdf983a9d618-elasticsearch-data\") pod \"elasticsearch-es-default-0\" (UID: \"3313dc1f-c9a1-4617-a9d6-cdf983a9d618\") " pod="service-telemetry/elasticsearch-es-default-0" Dec 03 00:21:34 crc kubenswrapper[4953]: I1203 00:21:34.735301 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"elasticsearch-logs\" (UniqueName: \"kubernetes.io/empty-dir/3313dc1f-c9a1-4617-a9d6-cdf983a9d618-elasticsearch-logs\") pod \"elasticsearch-es-default-0\" (UID: \"3313dc1f-c9a1-4617-a9d6-cdf983a9d618\") " pod="service-telemetry/elasticsearch-es-default-0" Dec 03 00:21:34 crc kubenswrapper[4953]: I1203 00:21:34.735282 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"elastic-internal-elasticsearch-bin-local\" (UniqueName: \"kubernetes.io/empty-dir/3313dc1f-c9a1-4617-a9d6-cdf983a9d618-elastic-internal-elasticsearch-bin-local\") pod \"elasticsearch-es-default-0\" (UID: \"3313dc1f-c9a1-4617-a9d6-cdf983a9d618\") " pod="service-telemetry/elasticsearch-es-default-0" Dec 03 00:21:34 crc kubenswrapper[4953]: I1203 00:21:34.735710 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"elastic-internal-elasticsearch-config-local\" (UniqueName: \"kubernetes.io/empty-dir/3313dc1f-c9a1-4617-a9d6-cdf983a9d618-elastic-internal-elasticsearch-config-local\") pod \"elasticsearch-es-default-0\" (UID: \"3313dc1f-c9a1-4617-a9d6-cdf983a9d618\") " pod="service-telemetry/elasticsearch-es-default-0" Dec 03 00:21:34 crc kubenswrapper[4953]: I1203 00:21:34.737193 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmp-volume\" (UniqueName: \"kubernetes.io/empty-dir/3313dc1f-c9a1-4617-a9d6-cdf983a9d618-tmp-volume\") pod \"elasticsearch-es-default-0\" (UID: \"3313dc1f-c9a1-4617-a9d6-cdf983a9d618\") " pod="service-telemetry/elasticsearch-es-default-0" Dec 03 00:21:34 crc kubenswrapper[4953]: I1203 00:21:34.737769 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"elastic-internal-elasticsearch-plugins-local\" (UniqueName: \"kubernetes.io/empty-dir/3313dc1f-c9a1-4617-a9d6-cdf983a9d618-elastic-internal-elasticsearch-plugins-local\") pod \"elasticsearch-es-default-0\" (UID: \"3313dc1f-c9a1-4617-a9d6-cdf983a9d618\") " pod="service-telemetry/elasticsearch-es-default-0" Dec 03 00:21:34 crc kubenswrapper[4953]: I1203 00:21:34.738136 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"elastic-internal-unicast-hosts\" (UniqueName: \"kubernetes.io/configmap/3313dc1f-c9a1-4617-a9d6-cdf983a9d618-elastic-internal-unicast-hosts\") pod \"elasticsearch-es-default-0\" (UID: \"3313dc1f-c9a1-4617-a9d6-cdf983a9d618\") " pod="service-telemetry/elasticsearch-es-default-0" Dec 03 00:21:34 crc kubenswrapper[4953]: I1203 00:21:34.738237 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"elastic-internal-scripts\" (UniqueName: \"kubernetes.io/configmap/3313dc1f-c9a1-4617-a9d6-cdf983a9d618-elastic-internal-scripts\") pod \"elasticsearch-es-default-0\" (UID: \"3313dc1f-c9a1-4617-a9d6-cdf983a9d618\") " pod="service-telemetry/elasticsearch-es-default-0" Dec 03 00:21:34 crc kubenswrapper[4953]: I1203 00:21:34.741693 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"elastic-internal-http-certificates\" (UniqueName: \"kubernetes.io/secret/3313dc1f-c9a1-4617-a9d6-cdf983a9d618-elastic-internal-http-certificates\") pod \"elasticsearch-es-default-0\" (UID: \"3313dc1f-c9a1-4617-a9d6-cdf983a9d618\") " pod="service-telemetry/elasticsearch-es-default-0" Dec 03 00:21:34 crc kubenswrapper[4953]: I1203 00:21:34.741817 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"elastic-internal-xpack-file-realm\" (UniqueName: \"kubernetes.io/secret/3313dc1f-c9a1-4617-a9d6-cdf983a9d618-elastic-internal-xpack-file-realm\") pod \"elasticsearch-es-default-0\" (UID: \"3313dc1f-c9a1-4617-a9d6-cdf983a9d618\") " pod="service-telemetry/elasticsearch-es-default-0" Dec 03 00:21:34 crc kubenswrapper[4953]: I1203 00:21:34.741883 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"downward-api\" (UniqueName: \"kubernetes.io/downward-api/3313dc1f-c9a1-4617-a9d6-cdf983a9d618-downward-api\") pod \"elasticsearch-es-default-0\" (UID: \"3313dc1f-c9a1-4617-a9d6-cdf983a9d618\") " pod="service-telemetry/elasticsearch-es-default-0" Dec 03 00:21:34 crc kubenswrapper[4953]: I1203 00:21:34.742517 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"elastic-internal-elasticsearch-config\" (UniqueName: \"kubernetes.io/secret/3313dc1f-c9a1-4617-a9d6-cdf983a9d618-elastic-internal-elasticsearch-config\") pod \"elasticsearch-es-default-0\" (UID: \"3313dc1f-c9a1-4617-a9d6-cdf983a9d618\") " pod="service-telemetry/elasticsearch-es-default-0" Dec 03 00:21:34 crc kubenswrapper[4953]: I1203 00:21:34.742810 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"elastic-internal-remote-certificate-authorities\" (UniqueName: \"kubernetes.io/secret/3313dc1f-c9a1-4617-a9d6-cdf983a9d618-elastic-internal-remote-certificate-authorities\") pod \"elasticsearch-es-default-0\" (UID: \"3313dc1f-c9a1-4617-a9d6-cdf983a9d618\") " pod="service-telemetry/elasticsearch-es-default-0" Dec 03 00:21:34 crc kubenswrapper[4953]: I1203 00:21:34.742914 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"elastic-internal-probe-user\" (UniqueName: \"kubernetes.io/secret/3313dc1f-c9a1-4617-a9d6-cdf983a9d618-elastic-internal-probe-user\") pod \"elasticsearch-es-default-0\" (UID: \"3313dc1f-c9a1-4617-a9d6-cdf983a9d618\") " pod="service-telemetry/elasticsearch-es-default-0" Dec 03 00:21:34 crc kubenswrapper[4953]: I1203 00:21:34.746674 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"elastic-internal-transport-certificates\" (UniqueName: \"kubernetes.io/secret/3313dc1f-c9a1-4617-a9d6-cdf983a9d618-elastic-internal-transport-certificates\") pod \"elasticsearch-es-default-0\" (UID: \"3313dc1f-c9a1-4617-a9d6-cdf983a9d618\") " pod="service-telemetry/elasticsearch-es-default-0" Dec 03 00:21:34 crc kubenswrapper[4953]: I1203 00:21:34.847155 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/elasticsearch-es-default-0" Dec 03 00:21:35 crc kubenswrapper[4953]: I1203 00:21:35.373945 4953 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931ahqk5h" Dec 03 00:21:35 crc kubenswrapper[4953]: I1203 00:21:35.415073 4953 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/elasticsearch-es-default-0"] Dec 03 00:21:35 crc kubenswrapper[4953]: W1203 00:21:35.431521 4953 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3313dc1f_c9a1_4617_a9d6_cdf983a9d618.slice/crio-f32a7da275038c15a4b01c11e3e119d909721bab53cb440dec2c4a8e1bab35ac WatchSource:0}: Error finding container f32a7da275038c15a4b01c11e3e119d909721bab53cb440dec2c4a8e1bab35ac: Status 404 returned error can't find the container with id f32a7da275038c15a4b01c11e3e119d909721bab53cb440dec2c4a8e1bab35ac Dec 03 00:21:35 crc kubenswrapper[4953]: I1203 00:21:35.452664 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/8bc0877b-282d-457a-a46a-c3202a31e2fb-bundle\") pod \"8bc0877b-282d-457a-a46a-c3202a31e2fb\" (UID: \"8bc0877b-282d-457a-a46a-c3202a31e2fb\") " Dec 03 00:21:35 crc kubenswrapper[4953]: I1203 00:21:35.452772 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/8bc0877b-282d-457a-a46a-c3202a31e2fb-util\") pod \"8bc0877b-282d-457a-a46a-c3202a31e2fb\" (UID: \"8bc0877b-282d-457a-a46a-c3202a31e2fb\") " Dec 03 00:21:35 crc kubenswrapper[4953]: I1203 00:21:35.452823 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-58gm2\" (UniqueName: \"kubernetes.io/projected/8bc0877b-282d-457a-a46a-c3202a31e2fb-kube-api-access-58gm2\") pod \"8bc0877b-282d-457a-a46a-c3202a31e2fb\" (UID: \"8bc0877b-282d-457a-a46a-c3202a31e2fb\") " Dec 03 00:21:35 crc kubenswrapper[4953]: I1203 00:21:35.454309 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8bc0877b-282d-457a-a46a-c3202a31e2fb-bundle" (OuterVolumeSpecName: "bundle") pod "8bc0877b-282d-457a-a46a-c3202a31e2fb" (UID: "8bc0877b-282d-457a-a46a-c3202a31e2fb"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 00:21:35 crc kubenswrapper[4953]: I1203 00:21:35.459281 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8bc0877b-282d-457a-a46a-c3202a31e2fb-kube-api-access-58gm2" (OuterVolumeSpecName: "kube-api-access-58gm2") pod "8bc0877b-282d-457a-a46a-c3202a31e2fb" (UID: "8bc0877b-282d-457a-a46a-c3202a31e2fb"). InnerVolumeSpecName "kube-api-access-58gm2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:21:35 crc kubenswrapper[4953]: I1203 00:21:35.465403 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8bc0877b-282d-457a-a46a-c3202a31e2fb-util" (OuterVolumeSpecName: "util") pod "8bc0877b-282d-457a-a46a-c3202a31e2fb" (UID: "8bc0877b-282d-457a-a46a-c3202a31e2fb"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 00:21:35 crc kubenswrapper[4953]: I1203 00:21:35.555102 4953 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/8bc0877b-282d-457a-a46a-c3202a31e2fb-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 00:21:35 crc kubenswrapper[4953]: I1203 00:21:35.555205 4953 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/8bc0877b-282d-457a-a46a-c3202a31e2fb-util\") on node \"crc\" DevicePath \"\"" Dec 03 00:21:35 crc kubenswrapper[4953]: I1203 00:21:35.555224 4953 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-58gm2\" (UniqueName: \"kubernetes.io/projected/8bc0877b-282d-457a-a46a-c3202a31e2fb-kube-api-access-58gm2\") on node \"crc\" DevicePath \"\"" Dec 03 00:21:35 crc kubenswrapper[4953]: I1203 00:21:35.809520 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/elasticsearch-es-default-0" event={"ID":"3313dc1f-c9a1-4617-a9d6-cdf983a9d618","Type":"ContainerStarted","Data":"f32a7da275038c15a4b01c11e3e119d909721bab53cb440dec2c4a8e1bab35ac"} Dec 03 00:21:35 crc kubenswrapper[4953]: I1203 00:21:35.811628 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931ahqk5h" event={"ID":"8bc0877b-282d-457a-a46a-c3202a31e2fb","Type":"ContainerDied","Data":"3bb493a46a0fcd601b96c0285efb54caa4e7a7cf5555f8d77a60711313c87ba8"} Dec 03 00:21:35 crc kubenswrapper[4953]: I1203 00:21:35.811659 4953 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3bb493a46a0fcd601b96c0285efb54caa4e7a7cf5555f8d77a60711313c87ba8" Dec 03 00:21:35 crc kubenswrapper[4953]: I1203 00:21:35.811799 4953 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931ahqk5h" Dec 03 00:21:35 crc kubenswrapper[4953]: I1203 00:21:35.813534 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-68bbd5c68d-pmttp" event={"ID":"fc13989b-f854-49fd-8901-3b206167aafc","Type":"ContainerStarted","Data":"a728c996f29e6ad1efc497e3ddfa0f858e292f557b7e10f5c62cea97982fc58b"} Dec 03 00:21:35 crc kubenswrapper[4953]: I1203 00:21:35.841331 4953 patch_prober.go:28] interesting pod/observability-operator-d8bb48f5d-64xzt container/operator namespace/openshift-operators: Readiness probe status=failure output="Get \"http://10.217.0.47:8081/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 03 00:21:35 crc kubenswrapper[4953]: I1203 00:21:35.841443 4953 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operators/observability-operator-d8bb48f5d-64xzt" podUID="1df642b8-586a-4bc5-8d9f-9f2cff7c6b80" containerName="operator" probeResult="failure" output="Get \"http://10.217.0.47:8081/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 03 00:21:35 crc kubenswrapper[4953]: I1203 00:21:35.852708 4953 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/obo-prometheus-operator-admission-webhook-68bbd5c68d-pmttp" podStartSLOduration=-9223371990.002102 podStartE2EDuration="46.852673186s" podCreationTimestamp="2025-12-03 00:20:49 +0000 UTC" firstStartedPulling="2025-12-03 00:20:50.913715649 +0000 UTC m=+839.897996432" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 00:21:35.83949819 +0000 UTC m=+884.823778973" watchObservedRunningTime="2025-12-03 00:21:35.852673186 +0000 UTC m=+884.836953969" Dec 03 00:21:39 crc kubenswrapper[4953]: I1203 00:21:39.699806 4953 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operators/observability-operator-d8bb48f5d-64xzt" Dec 03 00:21:40 crc kubenswrapper[4953]: I1203 00:21:40.272443 4953 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operators/perses-operator-5446b9c989-hqmj8" Dec 03 00:21:45 crc kubenswrapper[4953]: I1203 00:21:45.222507 4953 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["service-telemetry/service-telemetry-operator-1-build"] Dec 03 00:21:45 crc kubenswrapper[4953]: E1203 00:21:45.223448 4953 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8bc0877b-282d-457a-a46a-c3202a31e2fb" containerName="pull" Dec 03 00:21:45 crc kubenswrapper[4953]: I1203 00:21:45.223469 4953 state_mem.go:107] "Deleted CPUSet assignment" podUID="8bc0877b-282d-457a-a46a-c3202a31e2fb" containerName="pull" Dec 03 00:21:45 crc kubenswrapper[4953]: E1203 00:21:45.223492 4953 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8bc0877b-282d-457a-a46a-c3202a31e2fb" containerName="extract" Dec 03 00:21:45 crc kubenswrapper[4953]: I1203 00:21:45.223500 4953 state_mem.go:107] "Deleted CPUSet assignment" podUID="8bc0877b-282d-457a-a46a-c3202a31e2fb" containerName="extract" Dec 03 00:21:45 crc kubenswrapper[4953]: E1203 00:21:45.223513 4953 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8bc0877b-282d-457a-a46a-c3202a31e2fb" containerName="util" Dec 03 00:21:45 crc kubenswrapper[4953]: I1203 00:21:45.223522 4953 state_mem.go:107] "Deleted CPUSet assignment" podUID="8bc0877b-282d-457a-a46a-c3202a31e2fb" containerName="util" Dec 03 00:21:45 crc kubenswrapper[4953]: I1203 00:21:45.223689 4953 memory_manager.go:354] "RemoveStaleState removing state" podUID="8bc0877b-282d-457a-a46a-c3202a31e2fb" containerName="extract" Dec 03 00:21:45 crc kubenswrapper[4953]: I1203 00:21:45.224720 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/service-telemetry-operator-1-build" Dec 03 00:21:45 crc kubenswrapper[4953]: I1203 00:21:45.227907 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"service-telemetry-operator-1-global-ca" Dec 03 00:21:45 crc kubenswrapper[4953]: I1203 00:21:45.228057 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"service-telemetry-operator-1-sys-config" Dec 03 00:21:45 crc kubenswrapper[4953]: I1203 00:21:45.230335 4953 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"builder-dockercfg-9prwd" Dec 03 00:21:45 crc kubenswrapper[4953]: I1203 00:21:45.230383 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"service-telemetry-operator-1-ca" Dec 03 00:21:45 crc kubenswrapper[4953]: I1203 00:21:45.255168 4953 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/service-telemetry-operator-1-build"] Dec 03 00:21:45 crc kubenswrapper[4953]: I1203 00:21:45.271452 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"builder-dockercfg-9prwd-pull\" (UniqueName: \"kubernetes.io/secret/e1753922-98e9-49f9-8669-0b3efac1b78f-builder-dockercfg-9prwd-pull\") pod \"service-telemetry-operator-1-build\" (UID: \"e1753922-98e9-49f9-8669-0b3efac1b78f\") " pod="service-telemetry/service-telemetry-operator-1-build" Dec 03 00:21:45 crc kubenswrapper[4953]: I1203 00:21:45.271501 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/e1753922-98e9-49f9-8669-0b3efac1b78f-container-storage-run\") pod \"service-telemetry-operator-1-build\" (UID: \"e1753922-98e9-49f9-8669-0b3efac1b78f\") " pod="service-telemetry/service-telemetry-operator-1-build" Dec 03 00:21:45 crc kubenswrapper[4953]: I1203 00:21:45.271522 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9k4t4\" (UniqueName: \"kubernetes.io/projected/e1753922-98e9-49f9-8669-0b3efac1b78f-kube-api-access-9k4t4\") pod \"service-telemetry-operator-1-build\" (UID: \"e1753922-98e9-49f9-8669-0b3efac1b78f\") " pod="service-telemetry/service-telemetry-operator-1-build" Dec 03 00:21:45 crc kubenswrapper[4953]: I1203 00:21:45.271543 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/e1753922-98e9-49f9-8669-0b3efac1b78f-node-pullsecrets\") pod \"service-telemetry-operator-1-build\" (UID: \"e1753922-98e9-49f9-8669-0b3efac1b78f\") " pod="service-telemetry/service-telemetry-operator-1-build" Dec 03 00:21:45 crc kubenswrapper[4953]: I1203 00:21:45.271574 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/e1753922-98e9-49f9-8669-0b3efac1b78f-build-blob-cache\") pod \"service-telemetry-operator-1-build\" (UID: \"e1753922-98e9-49f9-8669-0b3efac1b78f\") " pod="service-telemetry/service-telemetry-operator-1-build" Dec 03 00:21:45 crc kubenswrapper[4953]: I1203 00:21:45.271592 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"builder-dockercfg-9prwd-push\" (UniqueName: \"kubernetes.io/secret/e1753922-98e9-49f9-8669-0b3efac1b78f-builder-dockercfg-9prwd-push\") pod \"service-telemetry-operator-1-build\" (UID: \"e1753922-98e9-49f9-8669-0b3efac1b78f\") " pod="service-telemetry/service-telemetry-operator-1-build" Dec 03 00:21:45 crc kubenswrapper[4953]: I1203 00:21:45.271611 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/e1753922-98e9-49f9-8669-0b3efac1b78f-build-system-configs\") pod \"service-telemetry-operator-1-build\" (UID: \"e1753922-98e9-49f9-8669-0b3efac1b78f\") " pod="service-telemetry/service-telemetry-operator-1-build" Dec 03 00:21:45 crc kubenswrapper[4953]: I1203 00:21:45.271631 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/e1753922-98e9-49f9-8669-0b3efac1b78f-container-storage-root\") pod \"service-telemetry-operator-1-build\" (UID: \"e1753922-98e9-49f9-8669-0b3efac1b78f\") " pod="service-telemetry/service-telemetry-operator-1-build" Dec 03 00:21:45 crc kubenswrapper[4953]: I1203 00:21:45.271659 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/e1753922-98e9-49f9-8669-0b3efac1b78f-buildworkdir\") pod \"service-telemetry-operator-1-build\" (UID: \"e1753922-98e9-49f9-8669-0b3efac1b78f\") " pod="service-telemetry/service-telemetry-operator-1-build" Dec 03 00:21:45 crc kubenswrapper[4953]: I1203 00:21:45.271681 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/e1753922-98e9-49f9-8669-0b3efac1b78f-buildcachedir\") pod \"service-telemetry-operator-1-build\" (UID: \"e1753922-98e9-49f9-8669-0b3efac1b78f\") " pod="service-telemetry/service-telemetry-operator-1-build" Dec 03 00:21:45 crc kubenswrapper[4953]: I1203 00:21:45.271695 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/e1753922-98e9-49f9-8669-0b3efac1b78f-build-proxy-ca-bundles\") pod \"service-telemetry-operator-1-build\" (UID: \"e1753922-98e9-49f9-8669-0b3efac1b78f\") " pod="service-telemetry/service-telemetry-operator-1-build" Dec 03 00:21:45 crc kubenswrapper[4953]: I1203 00:21:45.271711 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/e1753922-98e9-49f9-8669-0b3efac1b78f-build-ca-bundles\") pod \"service-telemetry-operator-1-build\" (UID: \"e1753922-98e9-49f9-8669-0b3efac1b78f\") " pod="service-telemetry/service-telemetry-operator-1-build" Dec 03 00:21:45 crc kubenswrapper[4953]: I1203 00:21:45.373607 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"builder-dockercfg-9prwd-pull\" (UniqueName: \"kubernetes.io/secret/e1753922-98e9-49f9-8669-0b3efac1b78f-builder-dockercfg-9prwd-pull\") pod \"service-telemetry-operator-1-build\" (UID: \"e1753922-98e9-49f9-8669-0b3efac1b78f\") " pod="service-telemetry/service-telemetry-operator-1-build" Dec 03 00:21:45 crc kubenswrapper[4953]: I1203 00:21:45.373657 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/e1753922-98e9-49f9-8669-0b3efac1b78f-container-storage-run\") pod \"service-telemetry-operator-1-build\" (UID: \"e1753922-98e9-49f9-8669-0b3efac1b78f\") " pod="service-telemetry/service-telemetry-operator-1-build" Dec 03 00:21:45 crc kubenswrapper[4953]: I1203 00:21:45.373678 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9k4t4\" (UniqueName: \"kubernetes.io/projected/e1753922-98e9-49f9-8669-0b3efac1b78f-kube-api-access-9k4t4\") pod \"service-telemetry-operator-1-build\" (UID: \"e1753922-98e9-49f9-8669-0b3efac1b78f\") " pod="service-telemetry/service-telemetry-operator-1-build" Dec 03 00:21:45 crc kubenswrapper[4953]: I1203 00:21:45.373703 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/e1753922-98e9-49f9-8669-0b3efac1b78f-node-pullsecrets\") pod \"service-telemetry-operator-1-build\" (UID: \"e1753922-98e9-49f9-8669-0b3efac1b78f\") " pod="service-telemetry/service-telemetry-operator-1-build" Dec 03 00:21:45 crc kubenswrapper[4953]: I1203 00:21:45.373738 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/e1753922-98e9-49f9-8669-0b3efac1b78f-build-blob-cache\") pod \"service-telemetry-operator-1-build\" (UID: \"e1753922-98e9-49f9-8669-0b3efac1b78f\") " pod="service-telemetry/service-telemetry-operator-1-build" Dec 03 00:21:45 crc kubenswrapper[4953]: I1203 00:21:45.373756 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"builder-dockercfg-9prwd-push\" (UniqueName: \"kubernetes.io/secret/e1753922-98e9-49f9-8669-0b3efac1b78f-builder-dockercfg-9prwd-push\") pod \"service-telemetry-operator-1-build\" (UID: \"e1753922-98e9-49f9-8669-0b3efac1b78f\") " pod="service-telemetry/service-telemetry-operator-1-build" Dec 03 00:21:45 crc kubenswrapper[4953]: I1203 00:21:45.373775 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/e1753922-98e9-49f9-8669-0b3efac1b78f-build-system-configs\") pod \"service-telemetry-operator-1-build\" (UID: \"e1753922-98e9-49f9-8669-0b3efac1b78f\") " pod="service-telemetry/service-telemetry-operator-1-build" Dec 03 00:21:45 crc kubenswrapper[4953]: I1203 00:21:45.373796 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/e1753922-98e9-49f9-8669-0b3efac1b78f-container-storage-root\") pod \"service-telemetry-operator-1-build\" (UID: \"e1753922-98e9-49f9-8669-0b3efac1b78f\") " pod="service-telemetry/service-telemetry-operator-1-build" Dec 03 00:21:45 crc kubenswrapper[4953]: I1203 00:21:45.373823 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/e1753922-98e9-49f9-8669-0b3efac1b78f-buildworkdir\") pod \"service-telemetry-operator-1-build\" (UID: \"e1753922-98e9-49f9-8669-0b3efac1b78f\") " pod="service-telemetry/service-telemetry-operator-1-build" Dec 03 00:21:45 crc kubenswrapper[4953]: I1203 00:21:45.373843 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/e1753922-98e9-49f9-8669-0b3efac1b78f-buildcachedir\") pod \"service-telemetry-operator-1-build\" (UID: \"e1753922-98e9-49f9-8669-0b3efac1b78f\") " pod="service-telemetry/service-telemetry-operator-1-build" Dec 03 00:21:45 crc kubenswrapper[4953]: I1203 00:21:45.373858 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/e1753922-98e9-49f9-8669-0b3efac1b78f-build-proxy-ca-bundles\") pod \"service-telemetry-operator-1-build\" (UID: \"e1753922-98e9-49f9-8669-0b3efac1b78f\") " pod="service-telemetry/service-telemetry-operator-1-build" Dec 03 00:21:45 crc kubenswrapper[4953]: I1203 00:21:45.373874 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/e1753922-98e9-49f9-8669-0b3efac1b78f-build-ca-bundles\") pod \"service-telemetry-operator-1-build\" (UID: \"e1753922-98e9-49f9-8669-0b3efac1b78f\") " pod="service-telemetry/service-telemetry-operator-1-build" Dec 03 00:21:45 crc kubenswrapper[4953]: I1203 00:21:45.374798 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/e1753922-98e9-49f9-8669-0b3efac1b78f-build-ca-bundles\") pod \"service-telemetry-operator-1-build\" (UID: \"e1753922-98e9-49f9-8669-0b3efac1b78f\") " pod="service-telemetry/service-telemetry-operator-1-build" Dec 03 00:21:45 crc kubenswrapper[4953]: I1203 00:21:45.378734 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/e1753922-98e9-49f9-8669-0b3efac1b78f-buildworkdir\") pod \"service-telemetry-operator-1-build\" (UID: \"e1753922-98e9-49f9-8669-0b3efac1b78f\") " pod="service-telemetry/service-telemetry-operator-1-build" Dec 03 00:21:45 crc kubenswrapper[4953]: I1203 00:21:45.379256 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/e1753922-98e9-49f9-8669-0b3efac1b78f-build-system-configs\") pod \"service-telemetry-operator-1-build\" (UID: \"e1753922-98e9-49f9-8669-0b3efac1b78f\") " pod="service-telemetry/service-telemetry-operator-1-build" Dec 03 00:21:45 crc kubenswrapper[4953]: I1203 00:21:45.379473 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/e1753922-98e9-49f9-8669-0b3efac1b78f-container-storage-root\") pod \"service-telemetry-operator-1-build\" (UID: \"e1753922-98e9-49f9-8669-0b3efac1b78f\") " pod="service-telemetry/service-telemetry-operator-1-build" Dec 03 00:21:45 crc kubenswrapper[4953]: I1203 00:21:45.379517 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/e1753922-98e9-49f9-8669-0b3efac1b78f-buildcachedir\") pod \"service-telemetry-operator-1-build\" (UID: \"e1753922-98e9-49f9-8669-0b3efac1b78f\") " pod="service-telemetry/service-telemetry-operator-1-build" Dec 03 00:21:45 crc kubenswrapper[4953]: I1203 00:21:45.380056 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/e1753922-98e9-49f9-8669-0b3efac1b78f-build-proxy-ca-bundles\") pod \"service-telemetry-operator-1-build\" (UID: \"e1753922-98e9-49f9-8669-0b3efac1b78f\") " pod="service-telemetry/service-telemetry-operator-1-build" Dec 03 00:21:45 crc kubenswrapper[4953]: I1203 00:21:45.380812 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/e1753922-98e9-49f9-8669-0b3efac1b78f-container-storage-run\") pod \"service-telemetry-operator-1-build\" (UID: \"e1753922-98e9-49f9-8669-0b3efac1b78f\") " pod="service-telemetry/service-telemetry-operator-1-build" Dec 03 00:21:45 crc kubenswrapper[4953]: I1203 00:21:45.380876 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/e1753922-98e9-49f9-8669-0b3efac1b78f-node-pullsecrets\") pod \"service-telemetry-operator-1-build\" (UID: \"e1753922-98e9-49f9-8669-0b3efac1b78f\") " pod="service-telemetry/service-telemetry-operator-1-build" Dec 03 00:21:45 crc kubenswrapper[4953]: I1203 00:21:45.387456 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"builder-dockercfg-9prwd-push\" (UniqueName: \"kubernetes.io/secret/e1753922-98e9-49f9-8669-0b3efac1b78f-builder-dockercfg-9prwd-push\") pod \"service-telemetry-operator-1-build\" (UID: \"e1753922-98e9-49f9-8669-0b3efac1b78f\") " pod="service-telemetry/service-telemetry-operator-1-build" Dec 03 00:21:45 crc kubenswrapper[4953]: I1203 00:21:45.389210 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"builder-dockercfg-9prwd-pull\" (UniqueName: \"kubernetes.io/secret/e1753922-98e9-49f9-8669-0b3efac1b78f-builder-dockercfg-9prwd-pull\") pod \"service-telemetry-operator-1-build\" (UID: \"e1753922-98e9-49f9-8669-0b3efac1b78f\") " pod="service-telemetry/service-telemetry-operator-1-build" Dec 03 00:21:45 crc kubenswrapper[4953]: I1203 00:21:45.397623 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/e1753922-98e9-49f9-8669-0b3efac1b78f-build-blob-cache\") pod \"service-telemetry-operator-1-build\" (UID: \"e1753922-98e9-49f9-8669-0b3efac1b78f\") " pod="service-telemetry/service-telemetry-operator-1-build" Dec 03 00:21:45 crc kubenswrapper[4953]: I1203 00:21:45.419667 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9k4t4\" (UniqueName: \"kubernetes.io/projected/e1753922-98e9-49f9-8669-0b3efac1b78f-kube-api-access-9k4t4\") pod \"service-telemetry-operator-1-build\" (UID: \"e1753922-98e9-49f9-8669-0b3efac1b78f\") " pod="service-telemetry/service-telemetry-operator-1-build" Dec 03 00:21:45 crc kubenswrapper[4953]: I1203 00:21:45.556508 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/service-telemetry-operator-1-build" Dec 03 00:21:48 crc kubenswrapper[4953]: I1203 00:21:48.474261 4953 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager-operator/cert-manager-operator-controller-manager-5446d6888b-mhgwl"] Dec 03 00:21:48 crc kubenswrapper[4953]: I1203 00:21:48.475871 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager-operator/cert-manager-operator-controller-manager-5446d6888b-mhgwl" Dec 03 00:21:48 crc kubenswrapper[4953]: I1203 00:21:48.485622 4953 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager-operator"/"cert-manager-operator-controller-manager-dockercfg-qxq66" Dec 03 00:21:48 crc kubenswrapper[4953]: I1203 00:21:48.485895 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager-operator"/"kube-root-ca.crt" Dec 03 00:21:48 crc kubenswrapper[4953]: I1203 00:21:48.486536 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager-operator"/"openshift-service-ca.crt" Dec 03 00:21:48 crc kubenswrapper[4953]: I1203 00:21:48.499389 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/d98b3207-d856-4464-a870-ea863266d32a-tmp\") pod \"cert-manager-operator-controller-manager-5446d6888b-mhgwl\" (UID: \"d98b3207-d856-4464-a870-ea863266d32a\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-5446d6888b-mhgwl" Dec 03 00:21:48 crc kubenswrapper[4953]: I1203 00:21:48.499526 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nrg4m\" (UniqueName: \"kubernetes.io/projected/d98b3207-d856-4464-a870-ea863266d32a-kube-api-access-nrg4m\") pod \"cert-manager-operator-controller-manager-5446d6888b-mhgwl\" (UID: \"d98b3207-d856-4464-a870-ea863266d32a\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-5446d6888b-mhgwl" Dec 03 00:21:48 crc kubenswrapper[4953]: I1203 00:21:48.511735 4953 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager-operator/cert-manager-operator-controller-manager-5446d6888b-mhgwl"] Dec 03 00:21:48 crc kubenswrapper[4953]: I1203 00:21:48.626830 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nrg4m\" (UniqueName: \"kubernetes.io/projected/d98b3207-d856-4464-a870-ea863266d32a-kube-api-access-nrg4m\") pod \"cert-manager-operator-controller-manager-5446d6888b-mhgwl\" (UID: \"d98b3207-d856-4464-a870-ea863266d32a\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-5446d6888b-mhgwl" Dec 03 00:21:48 crc kubenswrapper[4953]: I1203 00:21:48.626911 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/d98b3207-d856-4464-a870-ea863266d32a-tmp\") pod \"cert-manager-operator-controller-manager-5446d6888b-mhgwl\" (UID: \"d98b3207-d856-4464-a870-ea863266d32a\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-5446d6888b-mhgwl" Dec 03 00:21:48 crc kubenswrapper[4953]: I1203 00:21:48.627532 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/d98b3207-d856-4464-a870-ea863266d32a-tmp\") pod \"cert-manager-operator-controller-manager-5446d6888b-mhgwl\" (UID: \"d98b3207-d856-4464-a870-ea863266d32a\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-5446d6888b-mhgwl" Dec 03 00:21:48 crc kubenswrapper[4953]: I1203 00:21:48.902254 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nrg4m\" (UniqueName: \"kubernetes.io/projected/d98b3207-d856-4464-a870-ea863266d32a-kube-api-access-nrg4m\") pod \"cert-manager-operator-controller-manager-5446d6888b-mhgwl\" (UID: \"d98b3207-d856-4464-a870-ea863266d32a\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-5446d6888b-mhgwl" Dec 03 00:21:49 crc kubenswrapper[4953]: I1203 00:21:49.179164 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager-operator/cert-manager-operator-controller-manager-5446d6888b-mhgwl" Dec 03 00:21:55 crc kubenswrapper[4953]: I1203 00:21:55.441169 4953 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["service-telemetry/service-telemetry-operator-1-build"] Dec 03 00:21:57 crc kubenswrapper[4953]: I1203 00:21:57.335226 4953 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["service-telemetry/service-telemetry-operator-2-build"] Dec 03 00:21:57 crc kubenswrapper[4953]: I1203 00:21:57.337106 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/service-telemetry-operator-2-build" Dec 03 00:21:57 crc kubenswrapper[4953]: I1203 00:21:57.339988 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"service-telemetry-operator-2-global-ca" Dec 03 00:21:57 crc kubenswrapper[4953]: I1203 00:21:57.340339 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"service-telemetry-operator-2-ca" Dec 03 00:21:57 crc kubenswrapper[4953]: I1203 00:21:57.340486 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"service-telemetry-operator-2-sys-config" Dec 03 00:21:57 crc kubenswrapper[4953]: I1203 00:21:57.367449 4953 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/service-telemetry-operator-2-build"] Dec 03 00:21:57 crc kubenswrapper[4953]: I1203 00:21:57.455684 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4ljpw\" (UniqueName: \"kubernetes.io/projected/6519fd57-6635-41d7-b5e0-2fb21731574f-kube-api-access-4ljpw\") pod \"service-telemetry-operator-2-build\" (UID: \"6519fd57-6635-41d7-b5e0-2fb21731574f\") " pod="service-telemetry/service-telemetry-operator-2-build" Dec 03 00:21:57 crc kubenswrapper[4953]: I1203 00:21:57.455766 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/6519fd57-6635-41d7-b5e0-2fb21731574f-node-pullsecrets\") pod \"service-telemetry-operator-2-build\" (UID: \"6519fd57-6635-41d7-b5e0-2fb21731574f\") " pod="service-telemetry/service-telemetry-operator-2-build" Dec 03 00:21:57 crc kubenswrapper[4953]: I1203 00:21:57.455829 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/6519fd57-6635-41d7-b5e0-2fb21731574f-buildcachedir\") pod \"service-telemetry-operator-2-build\" (UID: \"6519fd57-6635-41d7-b5e0-2fb21731574f\") " pod="service-telemetry/service-telemetry-operator-2-build" Dec 03 00:21:57 crc kubenswrapper[4953]: I1203 00:21:57.455861 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/6519fd57-6635-41d7-b5e0-2fb21731574f-build-blob-cache\") pod \"service-telemetry-operator-2-build\" (UID: \"6519fd57-6635-41d7-b5e0-2fb21731574f\") " pod="service-telemetry/service-telemetry-operator-2-build" Dec 03 00:21:57 crc kubenswrapper[4953]: I1203 00:21:57.455909 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"builder-dockercfg-9prwd-push\" (UniqueName: \"kubernetes.io/secret/6519fd57-6635-41d7-b5e0-2fb21731574f-builder-dockercfg-9prwd-push\") pod \"service-telemetry-operator-2-build\" (UID: \"6519fd57-6635-41d7-b5e0-2fb21731574f\") " pod="service-telemetry/service-telemetry-operator-2-build" Dec 03 00:21:57 crc kubenswrapper[4953]: I1203 00:21:57.455943 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/6519fd57-6635-41d7-b5e0-2fb21731574f-build-proxy-ca-bundles\") pod \"service-telemetry-operator-2-build\" (UID: \"6519fd57-6635-41d7-b5e0-2fb21731574f\") " pod="service-telemetry/service-telemetry-operator-2-build" Dec 03 00:21:57 crc kubenswrapper[4953]: I1203 00:21:57.456287 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/6519fd57-6635-41d7-b5e0-2fb21731574f-buildworkdir\") pod \"service-telemetry-operator-2-build\" (UID: \"6519fd57-6635-41d7-b5e0-2fb21731574f\") " pod="service-telemetry/service-telemetry-operator-2-build" Dec 03 00:21:57 crc kubenswrapper[4953]: I1203 00:21:57.456389 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/6519fd57-6635-41d7-b5e0-2fb21731574f-container-storage-root\") pod \"service-telemetry-operator-2-build\" (UID: \"6519fd57-6635-41d7-b5e0-2fb21731574f\") " pod="service-telemetry/service-telemetry-operator-2-build" Dec 03 00:21:57 crc kubenswrapper[4953]: I1203 00:21:57.456509 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"builder-dockercfg-9prwd-pull\" (UniqueName: \"kubernetes.io/secret/6519fd57-6635-41d7-b5e0-2fb21731574f-builder-dockercfg-9prwd-pull\") pod \"service-telemetry-operator-2-build\" (UID: \"6519fd57-6635-41d7-b5e0-2fb21731574f\") " pod="service-telemetry/service-telemetry-operator-2-build" Dec 03 00:21:57 crc kubenswrapper[4953]: I1203 00:21:57.456697 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/6519fd57-6635-41d7-b5e0-2fb21731574f-build-ca-bundles\") pod \"service-telemetry-operator-2-build\" (UID: \"6519fd57-6635-41d7-b5e0-2fb21731574f\") " pod="service-telemetry/service-telemetry-operator-2-build" Dec 03 00:21:57 crc kubenswrapper[4953]: I1203 00:21:57.457498 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/6519fd57-6635-41d7-b5e0-2fb21731574f-build-system-configs\") pod \"service-telemetry-operator-2-build\" (UID: \"6519fd57-6635-41d7-b5e0-2fb21731574f\") " pod="service-telemetry/service-telemetry-operator-2-build" Dec 03 00:21:57 crc kubenswrapper[4953]: I1203 00:21:57.457583 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/6519fd57-6635-41d7-b5e0-2fb21731574f-container-storage-run\") pod \"service-telemetry-operator-2-build\" (UID: \"6519fd57-6635-41d7-b5e0-2fb21731574f\") " pod="service-telemetry/service-telemetry-operator-2-build" Dec 03 00:21:57 crc kubenswrapper[4953]: I1203 00:21:57.559329 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/6519fd57-6635-41d7-b5e0-2fb21731574f-buildworkdir\") pod \"service-telemetry-operator-2-build\" (UID: \"6519fd57-6635-41d7-b5e0-2fb21731574f\") " pod="service-telemetry/service-telemetry-operator-2-build" Dec 03 00:21:57 crc kubenswrapper[4953]: I1203 00:21:57.559424 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/6519fd57-6635-41d7-b5e0-2fb21731574f-container-storage-root\") pod \"service-telemetry-operator-2-build\" (UID: \"6519fd57-6635-41d7-b5e0-2fb21731574f\") " pod="service-telemetry/service-telemetry-operator-2-build" Dec 03 00:21:57 crc kubenswrapper[4953]: I1203 00:21:57.559456 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"builder-dockercfg-9prwd-pull\" (UniqueName: \"kubernetes.io/secret/6519fd57-6635-41d7-b5e0-2fb21731574f-builder-dockercfg-9prwd-pull\") pod \"service-telemetry-operator-2-build\" (UID: \"6519fd57-6635-41d7-b5e0-2fb21731574f\") " pod="service-telemetry/service-telemetry-operator-2-build" Dec 03 00:21:57 crc kubenswrapper[4953]: I1203 00:21:57.559479 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/6519fd57-6635-41d7-b5e0-2fb21731574f-build-ca-bundles\") pod \"service-telemetry-operator-2-build\" (UID: \"6519fd57-6635-41d7-b5e0-2fb21731574f\") " pod="service-telemetry/service-telemetry-operator-2-build" Dec 03 00:21:57 crc kubenswrapper[4953]: I1203 00:21:57.559565 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/6519fd57-6635-41d7-b5e0-2fb21731574f-build-system-configs\") pod \"service-telemetry-operator-2-build\" (UID: \"6519fd57-6635-41d7-b5e0-2fb21731574f\") " pod="service-telemetry/service-telemetry-operator-2-build" Dec 03 00:21:57 crc kubenswrapper[4953]: I1203 00:21:57.561237 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/6519fd57-6635-41d7-b5e0-2fb21731574f-container-storage-run\") pod \"service-telemetry-operator-2-build\" (UID: \"6519fd57-6635-41d7-b5e0-2fb21731574f\") " pod="service-telemetry/service-telemetry-operator-2-build" Dec 03 00:21:57 crc kubenswrapper[4953]: I1203 00:21:57.560051 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/6519fd57-6635-41d7-b5e0-2fb21731574f-container-storage-root\") pod \"service-telemetry-operator-2-build\" (UID: \"6519fd57-6635-41d7-b5e0-2fb21731574f\") " pod="service-telemetry/service-telemetry-operator-2-build" Dec 03 00:21:57 crc kubenswrapper[4953]: I1203 00:21:57.560384 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/6519fd57-6635-41d7-b5e0-2fb21731574f-build-system-configs\") pod \"service-telemetry-operator-2-build\" (UID: \"6519fd57-6635-41d7-b5e0-2fb21731574f\") " pod="service-telemetry/service-telemetry-operator-2-build" Dec 03 00:21:57 crc kubenswrapper[4953]: I1203 00:21:57.561662 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/6519fd57-6635-41d7-b5e0-2fb21731574f-container-storage-run\") pod \"service-telemetry-operator-2-build\" (UID: \"6519fd57-6635-41d7-b5e0-2fb21731574f\") " pod="service-telemetry/service-telemetry-operator-2-build" Dec 03 00:21:57 crc kubenswrapper[4953]: I1203 00:21:57.561174 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/6519fd57-6635-41d7-b5e0-2fb21731574f-build-ca-bundles\") pod \"service-telemetry-operator-2-build\" (UID: \"6519fd57-6635-41d7-b5e0-2fb21731574f\") " pod="service-telemetry/service-telemetry-operator-2-build" Dec 03 00:21:57 crc kubenswrapper[4953]: I1203 00:21:57.559903 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/6519fd57-6635-41d7-b5e0-2fb21731574f-buildworkdir\") pod \"service-telemetry-operator-2-build\" (UID: \"6519fd57-6635-41d7-b5e0-2fb21731574f\") " pod="service-telemetry/service-telemetry-operator-2-build" Dec 03 00:21:57 crc kubenswrapper[4953]: I1203 00:21:57.561754 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4ljpw\" (UniqueName: \"kubernetes.io/projected/6519fd57-6635-41d7-b5e0-2fb21731574f-kube-api-access-4ljpw\") pod \"service-telemetry-operator-2-build\" (UID: \"6519fd57-6635-41d7-b5e0-2fb21731574f\") " pod="service-telemetry/service-telemetry-operator-2-build" Dec 03 00:21:57 crc kubenswrapper[4953]: I1203 00:21:57.561786 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/6519fd57-6635-41d7-b5e0-2fb21731574f-node-pullsecrets\") pod \"service-telemetry-operator-2-build\" (UID: \"6519fd57-6635-41d7-b5e0-2fb21731574f\") " pod="service-telemetry/service-telemetry-operator-2-build" Dec 03 00:21:57 crc kubenswrapper[4953]: I1203 00:21:57.561812 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/6519fd57-6635-41d7-b5e0-2fb21731574f-buildcachedir\") pod \"service-telemetry-operator-2-build\" (UID: \"6519fd57-6635-41d7-b5e0-2fb21731574f\") " pod="service-telemetry/service-telemetry-operator-2-build" Dec 03 00:21:57 crc kubenswrapper[4953]: I1203 00:21:57.561834 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/6519fd57-6635-41d7-b5e0-2fb21731574f-build-blob-cache\") pod \"service-telemetry-operator-2-build\" (UID: \"6519fd57-6635-41d7-b5e0-2fb21731574f\") " pod="service-telemetry/service-telemetry-operator-2-build" Dec 03 00:21:57 crc kubenswrapper[4953]: I1203 00:21:57.561857 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"builder-dockercfg-9prwd-push\" (UniqueName: \"kubernetes.io/secret/6519fd57-6635-41d7-b5e0-2fb21731574f-builder-dockercfg-9prwd-push\") pod \"service-telemetry-operator-2-build\" (UID: \"6519fd57-6635-41d7-b5e0-2fb21731574f\") " pod="service-telemetry/service-telemetry-operator-2-build" Dec 03 00:21:57 crc kubenswrapper[4953]: I1203 00:21:57.561877 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/6519fd57-6635-41d7-b5e0-2fb21731574f-build-proxy-ca-bundles\") pod \"service-telemetry-operator-2-build\" (UID: \"6519fd57-6635-41d7-b5e0-2fb21731574f\") " pod="service-telemetry/service-telemetry-operator-2-build" Dec 03 00:21:57 crc kubenswrapper[4953]: I1203 00:21:57.562356 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/6519fd57-6635-41d7-b5e0-2fb21731574f-build-proxy-ca-bundles\") pod \"service-telemetry-operator-2-build\" (UID: \"6519fd57-6635-41d7-b5e0-2fb21731574f\") " pod="service-telemetry/service-telemetry-operator-2-build" Dec 03 00:21:57 crc kubenswrapper[4953]: I1203 00:21:57.562573 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/6519fd57-6635-41d7-b5e0-2fb21731574f-buildcachedir\") pod \"service-telemetry-operator-2-build\" (UID: \"6519fd57-6635-41d7-b5e0-2fb21731574f\") " pod="service-telemetry/service-telemetry-operator-2-build" Dec 03 00:21:57 crc kubenswrapper[4953]: I1203 00:21:57.562997 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/6519fd57-6635-41d7-b5e0-2fb21731574f-node-pullsecrets\") pod \"service-telemetry-operator-2-build\" (UID: \"6519fd57-6635-41d7-b5e0-2fb21731574f\") " pod="service-telemetry/service-telemetry-operator-2-build" Dec 03 00:21:57 crc kubenswrapper[4953]: I1203 00:21:57.563015 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/6519fd57-6635-41d7-b5e0-2fb21731574f-build-blob-cache\") pod \"service-telemetry-operator-2-build\" (UID: \"6519fd57-6635-41d7-b5e0-2fb21731574f\") " pod="service-telemetry/service-telemetry-operator-2-build" Dec 03 00:21:57 crc kubenswrapper[4953]: I1203 00:21:57.570805 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"builder-dockercfg-9prwd-push\" (UniqueName: \"kubernetes.io/secret/6519fd57-6635-41d7-b5e0-2fb21731574f-builder-dockercfg-9prwd-push\") pod \"service-telemetry-operator-2-build\" (UID: \"6519fd57-6635-41d7-b5e0-2fb21731574f\") " pod="service-telemetry/service-telemetry-operator-2-build" Dec 03 00:21:57 crc kubenswrapper[4953]: I1203 00:21:57.588525 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"builder-dockercfg-9prwd-pull\" (UniqueName: \"kubernetes.io/secret/6519fd57-6635-41d7-b5e0-2fb21731574f-builder-dockercfg-9prwd-pull\") pod \"service-telemetry-operator-2-build\" (UID: \"6519fd57-6635-41d7-b5e0-2fb21731574f\") " pod="service-telemetry/service-telemetry-operator-2-build" Dec 03 00:21:57 crc kubenswrapper[4953]: I1203 00:21:57.600149 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4ljpw\" (UniqueName: \"kubernetes.io/projected/6519fd57-6635-41d7-b5e0-2fb21731574f-kube-api-access-4ljpw\") pod \"service-telemetry-operator-2-build\" (UID: \"6519fd57-6635-41d7-b5e0-2fb21731574f\") " pod="service-telemetry/service-telemetry-operator-2-build" Dec 03 00:21:57 crc kubenswrapper[4953]: I1203 00:21:57.675132 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/service-telemetry-operator-2-build" Dec 03 00:22:02 crc kubenswrapper[4953]: I1203 00:22:02.226799 4953 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager-operator/cert-manager-operator-controller-manager-5446d6888b-mhgwl"] Dec 03 00:22:02 crc kubenswrapper[4953]: I1203 00:22:02.310500 4953 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["service-telemetry/service-telemetry-operator-1-build"] Dec 03 00:22:02 crc kubenswrapper[4953]: W1203 00:22:02.399260 4953 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode1753922_98e9_49f9_8669_0b3efac1b78f.slice/crio-e6d678481d618cf5ce0ffd3d09020fe2d4c66041ad90b1aa22b0fa577bfef6f4 WatchSource:0}: Error finding container e6d678481d618cf5ce0ffd3d09020fe2d4c66041ad90b1aa22b0fa577bfef6f4: Status 404 returned error can't find the container with id e6d678481d618cf5ce0ffd3d09020fe2d4c66041ad90b1aa22b0fa577bfef6f4 Dec 03 00:22:02 crc kubenswrapper[4953]: E1203 00:22:02.536627 4953 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="registry.connect.redhat.com/elastic/elasticsearch:7.17.20" Dec 03 00:22:02 crc kubenswrapper[4953]: E1203 00:22:02.537508 4953 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:elastic-internal-init-filesystem,Image:registry.connect.redhat.com/elastic/elasticsearch:7.17.20,Command:[bash -c /mnt/elastic-internal/scripts/prepare-fs.sh],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:POD_NAME,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.name,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:NODE_NAME,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:spec.nodeName,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:NAMESPACE,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:HEADLESS_SERVICE_NAME,Value:elasticsearch-es-default,ValueFrom:nil,},EnvVar{Name:PROBE_PASSWORD_PATH,Value:/mnt/elastic-internal/pod-mounted-users/elastic-internal-probe,ValueFrom:nil,},EnvVar{Name:PROBE_USERNAME,Value:elastic-internal-probe,ValueFrom:nil,},EnvVar{Name:READINESS_PROBE_PROTOCOL,Value:https,ValueFrom:nil,},EnvVar{Name:NSS_SDB_USE_CACHE,Value:no,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{100 -3} {} 100m DecimalSI},memory: {{52428800 0} {} 50Mi BinarySI},},Requests:ResourceList{cpu: {{100 -3} {} 100m DecimalSI},memory: {{52428800 0} {} 50Mi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:downward-api,ReadOnly:true,MountPath:/mnt/elastic-internal/downward-api,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:elastic-internal-elasticsearch-bin-local,ReadOnly:false,MountPath:/mnt/elastic-internal/elasticsearch-bin-local,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:elastic-internal-elasticsearch-config,ReadOnly:true,MountPath:/mnt/elastic-internal/elasticsearch-config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:elastic-internal-elasticsearch-config-local,ReadOnly:false,MountPath:/mnt/elastic-internal/elasticsearch-config-local,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:elastic-internal-elasticsearch-plugins-local,ReadOnly:false,MountPath:/mnt/elastic-internal/elasticsearch-plugins-local,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:elastic-internal-http-certificates,ReadOnly:true,MountPath:/usr/share/elasticsearch/config/http-certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:elastic-internal-probe-user,ReadOnly:true,MountPath:/mnt/elastic-internal/pod-mounted-users,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:elastic-internal-remote-certificate-authorities,ReadOnly:true,MountPath:/usr/share/elasticsearch/config/transport-remote-certs/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:elastic-internal-scripts,ReadOnly:true,MountPath:/mnt/elastic-internal/scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:elastic-internal-transport-certificates,ReadOnly:true,MountPath:/mnt/elastic-internal/transport-certificates,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:elastic-internal-unicast-hosts,ReadOnly:true,MountPath:/mnt/elastic-internal/unicast-hosts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:elastic-internal-xpack-file-realm,ReadOnly:true,MountPath:/mnt/elastic-internal/xpack-file-realm,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:elasticsearch-data,ReadOnly:false,MountPath:/usr/share/elasticsearch/data,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:elasticsearch-logs,ReadOnly:false,MountPath:/usr/share/elasticsearch/logs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:tmp-volume,ReadOnly:false,MountPath:/tmp,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*false,SELinuxOptions:nil,RunAsUser:*1000670000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:*true,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod elasticsearch-es-default-0_service-telemetry(3313dc1f-c9a1-4617-a9d6-cdf983a9d618): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 03 00:22:02 crc kubenswrapper[4953]: E1203 00:22:02.538765 4953 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"elastic-internal-init-filesystem\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="service-telemetry/elasticsearch-es-default-0" podUID="3313dc1f-c9a1-4617-a9d6-cdf983a9d618" Dec 03 00:22:02 crc kubenswrapper[4953]: I1203 00:22:02.629816 4953 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/service-telemetry-operator-2-build"] Dec 03 00:22:03 crc kubenswrapper[4953]: I1203 00:22:03.122020 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager-operator/cert-manager-operator-controller-manager-5446d6888b-mhgwl" event={"ID":"d98b3207-d856-4464-a870-ea863266d32a","Type":"ContainerStarted","Data":"5a2b47e6e4af748e92402a8de3788352534f30ccb2f6859612d13c1422b94491"} Dec 03 00:22:03 crc kubenswrapper[4953]: I1203 00:22:03.123430 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/interconnect-operator-5bb49f789d-q28d5" event={"ID":"4ae35630-6002-4c86-84ed-135aeda64570","Type":"ContainerStarted","Data":"287a4b69a7a5ffd7a4b4d3134e2b9b320c3e8d501fd8376665a466dda081cf53"} Dec 03 00:22:03 crc kubenswrapper[4953]: I1203 00:22:03.124656 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/service-telemetry-operator-2-build" event={"ID":"6519fd57-6635-41d7-b5e0-2fb21731574f","Type":"ContainerStarted","Data":"07c4a3a7f333bcf6a1530d2246b9cffd33d5bcae3f9c90803773c00e1c540078"} Dec 03 00:22:03 crc kubenswrapper[4953]: I1203 00:22:03.126132 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/service-telemetry-operator-1-build" event={"ID":"e1753922-98e9-49f9-8669-0b3efac1b78f","Type":"ContainerStarted","Data":"e6d678481d618cf5ce0ffd3d09020fe2d4c66041ad90b1aa22b0fa577bfef6f4"} Dec 03 00:22:03 crc kubenswrapper[4953]: E1203 00:22:03.128450 4953 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"elastic-internal-init-filesystem\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.connect.redhat.com/elastic/elasticsearch:7.17.20\\\"\"" pod="service-telemetry/elasticsearch-es-default-0" podUID="3313dc1f-c9a1-4617-a9d6-cdf983a9d618" Dec 03 00:22:03 crc kubenswrapper[4953]: I1203 00:22:03.147373 4953 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="service-telemetry/interconnect-operator-5bb49f789d-q28d5" podStartSLOduration=21.894655735 podStartE2EDuration="1m4.147343519s" podCreationTimestamp="2025-12-03 00:20:59 +0000 UTC" firstStartedPulling="2025-12-03 00:21:19.292687965 +0000 UTC m=+868.276968738" lastFinishedPulling="2025-12-03 00:22:01.545375739 +0000 UTC m=+910.529656522" observedRunningTime="2025-12-03 00:22:03.145439542 +0000 UTC m=+912.129720325" watchObservedRunningTime="2025-12-03 00:22:03.147343519 +0000 UTC m=+912.131624302" Dec 03 00:22:03 crc kubenswrapper[4953]: I1203 00:22:03.331059 4953 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/elasticsearch-es-default-0"] Dec 03 00:22:03 crc kubenswrapper[4953]: I1203 00:22:03.368708 4953 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/elasticsearch-es-default-0"] Dec 03 00:22:04 crc kubenswrapper[4953]: E1203 00:22:04.139229 4953 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"elastic-internal-init-filesystem\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.connect.redhat.com/elastic/elasticsearch:7.17.20\\\"\"" pod="service-telemetry/elasticsearch-es-default-0" podUID="3313dc1f-c9a1-4617-a9d6-cdf983a9d618" Dec 03 00:22:05 crc kubenswrapper[4953]: E1203 00:22:05.145989 4953 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"elastic-internal-init-filesystem\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.connect.redhat.com/elastic/elasticsearch:7.17.20\\\"\"" pod="service-telemetry/elasticsearch-es-default-0" podUID="3313dc1f-c9a1-4617-a9d6-cdf983a9d618" Dec 03 00:22:12 crc kubenswrapper[4953]: I1203 00:22:12.219124 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager-operator/cert-manager-operator-controller-manager-5446d6888b-mhgwl" event={"ID":"d98b3207-d856-4464-a870-ea863266d32a","Type":"ContainerStarted","Data":"49ecc021982c4f8d10590ca2756d32c206581d4bf6b4b39a76d816576d039d35"} Dec 03 00:22:12 crc kubenswrapper[4953]: I1203 00:22:12.244675 4953 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager-operator/cert-manager-operator-controller-manager-5446d6888b-mhgwl" podStartSLOduration=14.496893015 podStartE2EDuration="24.244646608s" podCreationTimestamp="2025-12-03 00:21:48 +0000 UTC" firstStartedPulling="2025-12-03 00:22:02.271713431 +0000 UTC m=+911.255994214" lastFinishedPulling="2025-12-03 00:22:12.019467024 +0000 UTC m=+921.003747807" observedRunningTime="2025-12-03 00:22:12.242744641 +0000 UTC m=+921.227025424" watchObservedRunningTime="2025-12-03 00:22:12.244646608 +0000 UTC m=+921.228927391" Dec 03 00:22:13 crc kubenswrapper[4953]: I1203 00:22:13.229921 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/service-telemetry-operator-1-build" event={"ID":"e1753922-98e9-49f9-8669-0b3efac1b78f","Type":"ContainerStarted","Data":"b2b34a16d20aea01c13d5f39c9a76f375efb0637864a50722bfd4e44853dd2ec"} Dec 03 00:22:13 crc kubenswrapper[4953]: I1203 00:22:13.230020 4953 kuberuntime_container.go:808] "Killing container with a grace period" pod="service-telemetry/service-telemetry-operator-1-build" podUID="e1753922-98e9-49f9-8669-0b3efac1b78f" containerName="manage-dockerfile" containerID="cri-o://b2b34a16d20aea01c13d5f39c9a76f375efb0637864a50722bfd4e44853dd2ec" gracePeriod=30 Dec 03 00:22:13 crc kubenswrapper[4953]: I1203 00:22:13.237553 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/service-telemetry-operator-2-build" event={"ID":"6519fd57-6635-41d7-b5e0-2fb21731574f","Type":"ContainerStarted","Data":"55e8f3ca2fc1d7ab9f8b5a1556bb62dd071ceed4d4d53ff8310ef628bf26598d"} Dec 03 00:22:13 crc kubenswrapper[4953]: I1203 00:22:13.926343 4953 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_service-telemetry-operator-1-build_e1753922-98e9-49f9-8669-0b3efac1b78f/manage-dockerfile/0.log" Dec 03 00:22:13 crc kubenswrapper[4953]: I1203 00:22:13.926920 4953 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/service-telemetry-operator-1-build" Dec 03 00:22:14 crc kubenswrapper[4953]: I1203 00:22:14.013509 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/e1753922-98e9-49f9-8669-0b3efac1b78f-build-ca-bundles\") pod \"e1753922-98e9-49f9-8669-0b3efac1b78f\" (UID: \"e1753922-98e9-49f9-8669-0b3efac1b78f\") " Dec 03 00:22:14 crc kubenswrapper[4953]: I1203 00:22:14.013573 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9k4t4\" (UniqueName: \"kubernetes.io/projected/e1753922-98e9-49f9-8669-0b3efac1b78f-kube-api-access-9k4t4\") pod \"e1753922-98e9-49f9-8669-0b3efac1b78f\" (UID: \"e1753922-98e9-49f9-8669-0b3efac1b78f\") " Dec 03 00:22:14 crc kubenswrapper[4953]: I1203 00:22:14.013601 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/e1753922-98e9-49f9-8669-0b3efac1b78f-container-storage-root\") pod \"e1753922-98e9-49f9-8669-0b3efac1b78f\" (UID: \"e1753922-98e9-49f9-8669-0b3efac1b78f\") " Dec 03 00:22:14 crc kubenswrapper[4953]: I1203 00:22:14.013671 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/e1753922-98e9-49f9-8669-0b3efac1b78f-buildcachedir\") pod \"e1753922-98e9-49f9-8669-0b3efac1b78f\" (UID: \"e1753922-98e9-49f9-8669-0b3efac1b78f\") " Dec 03 00:22:14 crc kubenswrapper[4953]: I1203 00:22:14.013756 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/e1753922-98e9-49f9-8669-0b3efac1b78f-node-pullsecrets\") pod \"e1753922-98e9-49f9-8669-0b3efac1b78f\" (UID: \"e1753922-98e9-49f9-8669-0b3efac1b78f\") " Dec 03 00:22:14 crc kubenswrapper[4953]: I1203 00:22:14.013785 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/e1753922-98e9-49f9-8669-0b3efac1b78f-build-system-configs\") pod \"e1753922-98e9-49f9-8669-0b3efac1b78f\" (UID: \"e1753922-98e9-49f9-8669-0b3efac1b78f\") " Dec 03 00:22:14 crc kubenswrapper[4953]: I1203 00:22:14.014111 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/e1753922-98e9-49f9-8669-0b3efac1b78f-container-storage-run\") pod \"e1753922-98e9-49f9-8669-0b3efac1b78f\" (UID: \"e1753922-98e9-49f9-8669-0b3efac1b78f\") " Dec 03 00:22:14 crc kubenswrapper[4953]: I1203 00:22:14.013836 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e1753922-98e9-49f9-8669-0b3efac1b78f-buildcachedir" (OuterVolumeSpecName: "buildcachedir") pod "e1753922-98e9-49f9-8669-0b3efac1b78f" (UID: "e1753922-98e9-49f9-8669-0b3efac1b78f"). InnerVolumeSpecName "buildcachedir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 00:22:14 crc kubenswrapper[4953]: I1203 00:22:14.013855 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e1753922-98e9-49f9-8669-0b3efac1b78f-node-pullsecrets" (OuterVolumeSpecName: "node-pullsecrets") pod "e1753922-98e9-49f9-8669-0b3efac1b78f" (UID: "e1753922-98e9-49f9-8669-0b3efac1b78f"). InnerVolumeSpecName "node-pullsecrets". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 00:22:14 crc kubenswrapper[4953]: I1203 00:22:14.014197 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e1753922-98e9-49f9-8669-0b3efac1b78f-container-storage-root" (OuterVolumeSpecName: "container-storage-root") pod "e1753922-98e9-49f9-8669-0b3efac1b78f" (UID: "e1753922-98e9-49f9-8669-0b3efac1b78f"). InnerVolumeSpecName "container-storage-root". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 00:22:14 crc kubenswrapper[4953]: I1203 00:22:14.014526 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e1753922-98e9-49f9-8669-0b3efac1b78f-container-storage-run" (OuterVolumeSpecName: "container-storage-run") pod "e1753922-98e9-49f9-8669-0b3efac1b78f" (UID: "e1753922-98e9-49f9-8669-0b3efac1b78f"). InnerVolumeSpecName "container-storage-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 00:22:14 crc kubenswrapper[4953]: I1203 00:22:14.014887 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e1753922-98e9-49f9-8669-0b3efac1b78f-build-system-configs" (OuterVolumeSpecName: "build-system-configs") pod "e1753922-98e9-49f9-8669-0b3efac1b78f" (UID: "e1753922-98e9-49f9-8669-0b3efac1b78f"). InnerVolumeSpecName "build-system-configs". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:22:14 crc kubenswrapper[4953]: I1203 00:22:14.014934 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e1753922-98e9-49f9-8669-0b3efac1b78f-build-ca-bundles" (OuterVolumeSpecName: "build-ca-bundles") pod "e1753922-98e9-49f9-8669-0b3efac1b78f" (UID: "e1753922-98e9-49f9-8669-0b3efac1b78f"). InnerVolumeSpecName "build-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:22:14 crc kubenswrapper[4953]: I1203 00:22:14.015422 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/e1753922-98e9-49f9-8669-0b3efac1b78f-build-blob-cache\") pod \"e1753922-98e9-49f9-8669-0b3efac1b78f\" (UID: \"e1753922-98e9-49f9-8669-0b3efac1b78f\") " Dec 03 00:22:14 crc kubenswrapper[4953]: I1203 00:22:14.015795 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/e1753922-98e9-49f9-8669-0b3efac1b78f-build-proxy-ca-bundles\") pod \"e1753922-98e9-49f9-8669-0b3efac1b78f\" (UID: \"e1753922-98e9-49f9-8669-0b3efac1b78f\") " Dec 03 00:22:14 crc kubenswrapper[4953]: I1203 00:22:14.015728 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e1753922-98e9-49f9-8669-0b3efac1b78f-build-blob-cache" (OuterVolumeSpecName: "build-blob-cache") pod "e1753922-98e9-49f9-8669-0b3efac1b78f" (UID: "e1753922-98e9-49f9-8669-0b3efac1b78f"). InnerVolumeSpecName "build-blob-cache". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 00:22:14 crc kubenswrapper[4953]: I1203 00:22:14.016106 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e1753922-98e9-49f9-8669-0b3efac1b78f-build-proxy-ca-bundles" (OuterVolumeSpecName: "build-proxy-ca-bundles") pod "e1753922-98e9-49f9-8669-0b3efac1b78f" (UID: "e1753922-98e9-49f9-8669-0b3efac1b78f"). InnerVolumeSpecName "build-proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:22:14 crc kubenswrapper[4953]: I1203 00:22:14.016262 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"builder-dockercfg-9prwd-push\" (UniqueName: \"kubernetes.io/secret/e1753922-98e9-49f9-8669-0b3efac1b78f-builder-dockercfg-9prwd-push\") pod \"e1753922-98e9-49f9-8669-0b3efac1b78f\" (UID: \"e1753922-98e9-49f9-8669-0b3efac1b78f\") " Dec 03 00:22:14 crc kubenswrapper[4953]: I1203 00:22:14.016595 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"builder-dockercfg-9prwd-pull\" (UniqueName: \"kubernetes.io/secret/e1753922-98e9-49f9-8669-0b3efac1b78f-builder-dockercfg-9prwd-pull\") pod \"e1753922-98e9-49f9-8669-0b3efac1b78f\" (UID: \"e1753922-98e9-49f9-8669-0b3efac1b78f\") " Dec 03 00:22:14 crc kubenswrapper[4953]: I1203 00:22:14.016680 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/e1753922-98e9-49f9-8669-0b3efac1b78f-buildworkdir\") pod \"e1753922-98e9-49f9-8669-0b3efac1b78f\" (UID: \"e1753922-98e9-49f9-8669-0b3efac1b78f\") " Dec 03 00:22:14 crc kubenswrapper[4953]: I1203 00:22:14.017155 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e1753922-98e9-49f9-8669-0b3efac1b78f-buildworkdir" (OuterVolumeSpecName: "buildworkdir") pod "e1753922-98e9-49f9-8669-0b3efac1b78f" (UID: "e1753922-98e9-49f9-8669-0b3efac1b78f"). InnerVolumeSpecName "buildworkdir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 00:22:14 crc kubenswrapper[4953]: I1203 00:22:14.017323 4953 reconciler_common.go:293] "Volume detached for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/e1753922-98e9-49f9-8669-0b3efac1b78f-build-ca-bundles\") on node \"crc\" DevicePath \"\"" Dec 03 00:22:14 crc kubenswrapper[4953]: I1203 00:22:14.017362 4953 reconciler_common.go:293] "Volume detached for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/e1753922-98e9-49f9-8669-0b3efac1b78f-container-storage-root\") on node \"crc\" DevicePath \"\"" Dec 03 00:22:14 crc kubenswrapper[4953]: I1203 00:22:14.017374 4953 reconciler_common.go:293] "Volume detached for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/e1753922-98e9-49f9-8669-0b3efac1b78f-buildcachedir\") on node \"crc\" DevicePath \"\"" Dec 03 00:22:14 crc kubenswrapper[4953]: I1203 00:22:14.017384 4953 reconciler_common.go:293] "Volume detached for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/e1753922-98e9-49f9-8669-0b3efac1b78f-node-pullsecrets\") on node \"crc\" DevicePath \"\"" Dec 03 00:22:14 crc kubenswrapper[4953]: I1203 00:22:14.017394 4953 reconciler_common.go:293] "Volume detached for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/e1753922-98e9-49f9-8669-0b3efac1b78f-build-system-configs\") on node \"crc\" DevicePath \"\"" Dec 03 00:22:14 crc kubenswrapper[4953]: I1203 00:22:14.017404 4953 reconciler_common.go:293] "Volume detached for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/e1753922-98e9-49f9-8669-0b3efac1b78f-container-storage-run\") on node \"crc\" DevicePath \"\"" Dec 03 00:22:14 crc kubenswrapper[4953]: I1203 00:22:14.017413 4953 reconciler_common.go:293] "Volume detached for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/e1753922-98e9-49f9-8669-0b3efac1b78f-build-blob-cache\") on node \"crc\" DevicePath \"\"" Dec 03 00:22:14 crc kubenswrapper[4953]: I1203 00:22:14.017422 4953 reconciler_common.go:293] "Volume detached for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/e1753922-98e9-49f9-8669-0b3efac1b78f-build-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Dec 03 00:22:14 crc kubenswrapper[4953]: I1203 00:22:14.017431 4953 reconciler_common.go:293] "Volume detached for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/e1753922-98e9-49f9-8669-0b3efac1b78f-buildworkdir\") on node \"crc\" DevicePath \"\"" Dec 03 00:22:14 crc kubenswrapper[4953]: I1203 00:22:14.026376 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e1753922-98e9-49f9-8669-0b3efac1b78f-builder-dockercfg-9prwd-pull" (OuterVolumeSpecName: "builder-dockercfg-9prwd-pull") pod "e1753922-98e9-49f9-8669-0b3efac1b78f" (UID: "e1753922-98e9-49f9-8669-0b3efac1b78f"). InnerVolumeSpecName "builder-dockercfg-9prwd-pull". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:22:14 crc kubenswrapper[4953]: I1203 00:22:14.040640 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e1753922-98e9-49f9-8669-0b3efac1b78f-builder-dockercfg-9prwd-push" (OuterVolumeSpecName: "builder-dockercfg-9prwd-push") pod "e1753922-98e9-49f9-8669-0b3efac1b78f" (UID: "e1753922-98e9-49f9-8669-0b3efac1b78f"). InnerVolumeSpecName "builder-dockercfg-9prwd-push". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:22:14 crc kubenswrapper[4953]: I1203 00:22:14.044300 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e1753922-98e9-49f9-8669-0b3efac1b78f-kube-api-access-9k4t4" (OuterVolumeSpecName: "kube-api-access-9k4t4") pod "e1753922-98e9-49f9-8669-0b3efac1b78f" (UID: "e1753922-98e9-49f9-8669-0b3efac1b78f"). InnerVolumeSpecName "kube-api-access-9k4t4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:22:14 crc kubenswrapper[4953]: I1203 00:22:14.119369 4953 reconciler_common.go:293] "Volume detached for volume \"builder-dockercfg-9prwd-push\" (UniqueName: \"kubernetes.io/secret/e1753922-98e9-49f9-8669-0b3efac1b78f-builder-dockercfg-9prwd-push\") on node \"crc\" DevicePath \"\"" Dec 03 00:22:14 crc kubenswrapper[4953]: I1203 00:22:14.119428 4953 reconciler_common.go:293] "Volume detached for volume \"builder-dockercfg-9prwd-pull\" (UniqueName: \"kubernetes.io/secret/e1753922-98e9-49f9-8669-0b3efac1b78f-builder-dockercfg-9prwd-pull\") on node \"crc\" DevicePath \"\"" Dec 03 00:22:14 crc kubenswrapper[4953]: I1203 00:22:14.119445 4953 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9k4t4\" (UniqueName: \"kubernetes.io/projected/e1753922-98e9-49f9-8669-0b3efac1b78f-kube-api-access-9k4t4\") on node \"crc\" DevicePath \"\"" Dec 03 00:22:14 crc kubenswrapper[4953]: I1203 00:22:14.247037 4953 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_service-telemetry-operator-1-build_e1753922-98e9-49f9-8669-0b3efac1b78f/manage-dockerfile/0.log" Dec 03 00:22:14 crc kubenswrapper[4953]: I1203 00:22:14.247562 4953 generic.go:334] "Generic (PLEG): container finished" podID="e1753922-98e9-49f9-8669-0b3efac1b78f" containerID="b2b34a16d20aea01c13d5f39c9a76f375efb0637864a50722bfd4e44853dd2ec" exitCode=1 Dec 03 00:22:14 crc kubenswrapper[4953]: I1203 00:22:14.247675 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/service-telemetry-operator-1-build" event={"ID":"e1753922-98e9-49f9-8669-0b3efac1b78f","Type":"ContainerDied","Data":"b2b34a16d20aea01c13d5f39c9a76f375efb0637864a50722bfd4e44853dd2ec"} Dec 03 00:22:14 crc kubenswrapper[4953]: I1203 00:22:14.247770 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/service-telemetry-operator-1-build" event={"ID":"e1753922-98e9-49f9-8669-0b3efac1b78f","Type":"ContainerDied","Data":"e6d678481d618cf5ce0ffd3d09020fe2d4c66041ad90b1aa22b0fa577bfef6f4"} Dec 03 00:22:14 crc kubenswrapper[4953]: I1203 00:22:14.247799 4953 scope.go:117] "RemoveContainer" containerID="b2b34a16d20aea01c13d5f39c9a76f375efb0637864a50722bfd4e44853dd2ec" Dec 03 00:22:14 crc kubenswrapper[4953]: I1203 00:22:14.248129 4953 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/service-telemetry-operator-1-build" Dec 03 00:22:14 crc kubenswrapper[4953]: I1203 00:22:14.295100 4953 scope.go:117] "RemoveContainer" containerID="b2b34a16d20aea01c13d5f39c9a76f375efb0637864a50722bfd4e44853dd2ec" Dec 03 00:22:14 crc kubenswrapper[4953]: E1203 00:22:14.297822 4953 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b2b34a16d20aea01c13d5f39c9a76f375efb0637864a50722bfd4e44853dd2ec\": container with ID starting with b2b34a16d20aea01c13d5f39c9a76f375efb0637864a50722bfd4e44853dd2ec not found: ID does not exist" containerID="b2b34a16d20aea01c13d5f39c9a76f375efb0637864a50722bfd4e44853dd2ec" Dec 03 00:22:14 crc kubenswrapper[4953]: I1203 00:22:14.297895 4953 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b2b34a16d20aea01c13d5f39c9a76f375efb0637864a50722bfd4e44853dd2ec"} err="failed to get container status \"b2b34a16d20aea01c13d5f39c9a76f375efb0637864a50722bfd4e44853dd2ec\": rpc error: code = NotFound desc = could not find container \"b2b34a16d20aea01c13d5f39c9a76f375efb0637864a50722bfd4e44853dd2ec\": container with ID starting with b2b34a16d20aea01c13d5f39c9a76f375efb0637864a50722bfd4e44853dd2ec not found: ID does not exist" Dec 03 00:22:14 crc kubenswrapper[4953]: I1203 00:22:14.370651 4953 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["service-telemetry/service-telemetry-operator-1-build"] Dec 03 00:22:14 crc kubenswrapper[4953]: I1203 00:22:14.383560 4953 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["service-telemetry/service-telemetry-operator-1-build"] Dec 03 00:22:15 crc kubenswrapper[4953]: I1203 00:22:15.356690 4953 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e1753922-98e9-49f9-8669-0b3efac1b78f" path="/var/lib/kubelet/pods/e1753922-98e9-49f9-8669-0b3efac1b78f/volumes" Dec 03 00:22:16 crc kubenswrapper[4953]: I1203 00:22:16.317673 4953 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-webhook-f4fb5df64-cg7mx"] Dec 03 00:22:16 crc kubenswrapper[4953]: E1203 00:22:16.318459 4953 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e1753922-98e9-49f9-8669-0b3efac1b78f" containerName="manage-dockerfile" Dec 03 00:22:16 crc kubenswrapper[4953]: I1203 00:22:16.318607 4953 state_mem.go:107] "Deleted CPUSet assignment" podUID="e1753922-98e9-49f9-8669-0b3efac1b78f" containerName="manage-dockerfile" Dec 03 00:22:16 crc kubenswrapper[4953]: I1203 00:22:16.318865 4953 memory_manager.go:354] "RemoveStaleState removing state" podUID="e1753922-98e9-49f9-8669-0b3efac1b78f" containerName="manage-dockerfile" Dec 03 00:22:16 crc kubenswrapper[4953]: I1203 00:22:16.319675 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-f4fb5df64-cg7mx" Dec 03 00:22:16 crc kubenswrapper[4953]: I1203 00:22:16.322426 4953 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-webhook-dockercfg-hhlg8" Dec 03 00:22:16 crc kubenswrapper[4953]: I1203 00:22:16.322602 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"kube-root-ca.crt" Dec 03 00:22:16 crc kubenswrapper[4953]: I1203 00:22:16.323653 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"openshift-service-ca.crt" Dec 03 00:22:16 crc kubenswrapper[4953]: I1203 00:22:16.335612 4953 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-f4fb5df64-cg7mx"] Dec 03 00:22:16 crc kubenswrapper[4953]: I1203 00:22:16.498341 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/37a3ced5-065d-4be8-a3e9-54deee67aefa-bound-sa-token\") pod \"cert-manager-webhook-f4fb5df64-cg7mx\" (UID: \"37a3ced5-065d-4be8-a3e9-54deee67aefa\") " pod="cert-manager/cert-manager-webhook-f4fb5df64-cg7mx" Dec 03 00:22:16 crc kubenswrapper[4953]: I1203 00:22:16.498396 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5ph8k\" (UniqueName: \"kubernetes.io/projected/37a3ced5-065d-4be8-a3e9-54deee67aefa-kube-api-access-5ph8k\") pod \"cert-manager-webhook-f4fb5df64-cg7mx\" (UID: \"37a3ced5-065d-4be8-a3e9-54deee67aefa\") " pod="cert-manager/cert-manager-webhook-f4fb5df64-cg7mx" Dec 03 00:22:16 crc kubenswrapper[4953]: I1203 00:22:16.600176 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/37a3ced5-065d-4be8-a3e9-54deee67aefa-bound-sa-token\") pod \"cert-manager-webhook-f4fb5df64-cg7mx\" (UID: \"37a3ced5-065d-4be8-a3e9-54deee67aefa\") " pod="cert-manager/cert-manager-webhook-f4fb5df64-cg7mx" Dec 03 00:22:16 crc kubenswrapper[4953]: I1203 00:22:16.600230 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5ph8k\" (UniqueName: \"kubernetes.io/projected/37a3ced5-065d-4be8-a3e9-54deee67aefa-kube-api-access-5ph8k\") pod \"cert-manager-webhook-f4fb5df64-cg7mx\" (UID: \"37a3ced5-065d-4be8-a3e9-54deee67aefa\") " pod="cert-manager/cert-manager-webhook-f4fb5df64-cg7mx" Dec 03 00:22:16 crc kubenswrapper[4953]: I1203 00:22:16.625037 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/37a3ced5-065d-4be8-a3e9-54deee67aefa-bound-sa-token\") pod \"cert-manager-webhook-f4fb5df64-cg7mx\" (UID: \"37a3ced5-065d-4be8-a3e9-54deee67aefa\") " pod="cert-manager/cert-manager-webhook-f4fb5df64-cg7mx" Dec 03 00:22:16 crc kubenswrapper[4953]: I1203 00:22:16.638265 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5ph8k\" (UniqueName: \"kubernetes.io/projected/37a3ced5-065d-4be8-a3e9-54deee67aefa-kube-api-access-5ph8k\") pod \"cert-manager-webhook-f4fb5df64-cg7mx\" (UID: \"37a3ced5-065d-4be8-a3e9-54deee67aefa\") " pod="cert-manager/cert-manager-webhook-f4fb5df64-cg7mx" Dec 03 00:22:16 crc kubenswrapper[4953]: I1203 00:22:16.638781 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-f4fb5df64-cg7mx" Dec 03 00:22:17 crc kubenswrapper[4953]: I1203 00:22:17.648714 4953 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-f4fb5df64-cg7mx"] Dec 03 00:22:18 crc kubenswrapper[4953]: I1203 00:22:18.279613 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-f4fb5df64-cg7mx" event={"ID":"37a3ced5-065d-4be8-a3e9-54deee67aefa","Type":"ContainerStarted","Data":"dee9fcbdb8b3c778b86bcfb7a0f950dbe2d0c15f606d098b5952d6dfc5828d78"} Dec 03 00:22:20 crc kubenswrapper[4953]: I1203 00:22:20.078735 4953 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-cainjector-855d9ccff4-gqhrh"] Dec 03 00:22:20 crc kubenswrapper[4953]: I1203 00:22:20.081595 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-855d9ccff4-gqhrh" Dec 03 00:22:20 crc kubenswrapper[4953]: I1203 00:22:20.085600 4953 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-cainjector-dockercfg-jqpkf" Dec 03 00:22:20 crc kubenswrapper[4953]: I1203 00:22:20.100584 4953 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-855d9ccff4-gqhrh"] Dec 03 00:22:20 crc kubenswrapper[4953]: I1203 00:22:20.168788 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/e2c5f086-eaf3-49ae-b67c-e3eb601c2c58-bound-sa-token\") pod \"cert-manager-cainjector-855d9ccff4-gqhrh\" (UID: \"e2c5f086-eaf3-49ae-b67c-e3eb601c2c58\") " pod="cert-manager/cert-manager-cainjector-855d9ccff4-gqhrh" Dec 03 00:22:20 crc kubenswrapper[4953]: I1203 00:22:20.168896 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2576h\" (UniqueName: \"kubernetes.io/projected/e2c5f086-eaf3-49ae-b67c-e3eb601c2c58-kube-api-access-2576h\") pod \"cert-manager-cainjector-855d9ccff4-gqhrh\" (UID: \"e2c5f086-eaf3-49ae-b67c-e3eb601c2c58\") " pod="cert-manager/cert-manager-cainjector-855d9ccff4-gqhrh" Dec 03 00:22:20 crc kubenswrapper[4953]: I1203 00:22:20.270964 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/e2c5f086-eaf3-49ae-b67c-e3eb601c2c58-bound-sa-token\") pod \"cert-manager-cainjector-855d9ccff4-gqhrh\" (UID: \"e2c5f086-eaf3-49ae-b67c-e3eb601c2c58\") " pod="cert-manager/cert-manager-cainjector-855d9ccff4-gqhrh" Dec 03 00:22:20 crc kubenswrapper[4953]: I1203 00:22:20.271280 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2576h\" (UniqueName: \"kubernetes.io/projected/e2c5f086-eaf3-49ae-b67c-e3eb601c2c58-kube-api-access-2576h\") pod \"cert-manager-cainjector-855d9ccff4-gqhrh\" (UID: \"e2c5f086-eaf3-49ae-b67c-e3eb601c2c58\") " pod="cert-manager/cert-manager-cainjector-855d9ccff4-gqhrh" Dec 03 00:22:20 crc kubenswrapper[4953]: I1203 00:22:20.293715 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2576h\" (UniqueName: \"kubernetes.io/projected/e2c5f086-eaf3-49ae-b67c-e3eb601c2c58-kube-api-access-2576h\") pod \"cert-manager-cainjector-855d9ccff4-gqhrh\" (UID: \"e2c5f086-eaf3-49ae-b67c-e3eb601c2c58\") " pod="cert-manager/cert-manager-cainjector-855d9ccff4-gqhrh" Dec 03 00:22:20 crc kubenswrapper[4953]: I1203 00:22:20.304758 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/e2c5f086-eaf3-49ae-b67c-e3eb601c2c58-bound-sa-token\") pod \"cert-manager-cainjector-855d9ccff4-gqhrh\" (UID: \"e2c5f086-eaf3-49ae-b67c-e3eb601c2c58\") " pod="cert-manager/cert-manager-cainjector-855d9ccff4-gqhrh" Dec 03 00:22:20 crc kubenswrapper[4953]: I1203 00:22:20.407187 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-855d9ccff4-gqhrh" Dec 03 00:22:21 crc kubenswrapper[4953]: I1203 00:22:21.034633 4953 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-855d9ccff4-gqhrh"] Dec 03 00:22:21 crc kubenswrapper[4953]: W1203 00:22:21.050893 4953 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode2c5f086_eaf3_49ae_b67c_e3eb601c2c58.slice/crio-13ba4413ca8db6e2e001f97664cad74456152ae8532bd2de348cb571648c0a05 WatchSource:0}: Error finding container 13ba4413ca8db6e2e001f97664cad74456152ae8532bd2de348cb571648c0a05: Status 404 returned error can't find the container with id 13ba4413ca8db6e2e001f97664cad74456152ae8532bd2de348cb571648c0a05 Dec 03 00:22:21 crc kubenswrapper[4953]: I1203 00:22:21.318186 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-855d9ccff4-gqhrh" event={"ID":"e2c5f086-eaf3-49ae-b67c-e3eb601c2c58","Type":"ContainerStarted","Data":"13ba4413ca8db6e2e001f97664cad74456152ae8532bd2de348cb571648c0a05"} Dec 03 00:22:21 crc kubenswrapper[4953]: I1203 00:22:21.329498 4953 generic.go:334] "Generic (PLEG): container finished" podID="6519fd57-6635-41d7-b5e0-2fb21731574f" containerID="55e8f3ca2fc1d7ab9f8b5a1556bb62dd071ceed4d4d53ff8310ef628bf26598d" exitCode=0 Dec 03 00:22:21 crc kubenswrapper[4953]: I1203 00:22:21.329568 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/service-telemetry-operator-2-build" event={"ID":"6519fd57-6635-41d7-b5e0-2fb21731574f","Type":"ContainerDied","Data":"55e8f3ca2fc1d7ab9f8b5a1556bb62dd071ceed4d4d53ff8310ef628bf26598d"} Dec 03 00:22:22 crc kubenswrapper[4953]: I1203 00:22:22.352898 4953 generic.go:334] "Generic (PLEG): container finished" podID="6519fd57-6635-41d7-b5e0-2fb21731574f" containerID="393742a679902fc395452a039daf7373f1b6d149f94de45c8326a55b660be24b" exitCode=0 Dec 03 00:22:22 crc kubenswrapper[4953]: I1203 00:22:22.353206 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/service-telemetry-operator-2-build" event={"ID":"6519fd57-6635-41d7-b5e0-2fb21731574f","Type":"ContainerDied","Data":"393742a679902fc395452a039daf7373f1b6d149f94de45c8326a55b660be24b"} Dec 03 00:22:22 crc kubenswrapper[4953]: I1203 00:22:22.398916 4953 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_service-telemetry-operator-2-build_6519fd57-6635-41d7-b5e0-2fb21731574f/manage-dockerfile/0.log" Dec 03 00:22:23 crc kubenswrapper[4953]: I1203 00:22:23.393329 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/elasticsearch-es-default-0" event={"ID":"3313dc1f-c9a1-4617-a9d6-cdf983a9d618","Type":"ContainerStarted","Data":"ceffa48b195f4207fcf6d3e94fdd111e3a4a18b49d939c9bb514560b02a6feab"} Dec 03 00:22:24 crc kubenswrapper[4953]: I1203 00:22:24.570426 4953 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-86cb77c54b-hq8kc"] Dec 03 00:22:24 crc kubenswrapper[4953]: I1203 00:22:24.572083 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-86cb77c54b-hq8kc" Dec 03 00:22:24 crc kubenswrapper[4953]: I1203 00:22:24.574413 4953 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-dockercfg-54vd4" Dec 03 00:22:24 crc kubenswrapper[4953]: I1203 00:22:24.588333 4953 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-86cb77c54b-hq8kc"] Dec 03 00:22:24 crc kubenswrapper[4953]: I1203 00:22:24.702066 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rv25m\" (UniqueName: \"kubernetes.io/projected/d6305126-b142-45e0-a058-1356b122cecd-kube-api-access-rv25m\") pod \"cert-manager-86cb77c54b-hq8kc\" (UID: \"d6305126-b142-45e0-a058-1356b122cecd\") " pod="cert-manager/cert-manager-86cb77c54b-hq8kc" Dec 03 00:22:24 crc kubenswrapper[4953]: I1203 00:22:24.702161 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/d6305126-b142-45e0-a058-1356b122cecd-bound-sa-token\") pod \"cert-manager-86cb77c54b-hq8kc\" (UID: \"d6305126-b142-45e0-a058-1356b122cecd\") " pod="cert-manager/cert-manager-86cb77c54b-hq8kc" Dec 03 00:22:24 crc kubenswrapper[4953]: I1203 00:22:24.803948 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/d6305126-b142-45e0-a058-1356b122cecd-bound-sa-token\") pod \"cert-manager-86cb77c54b-hq8kc\" (UID: \"d6305126-b142-45e0-a058-1356b122cecd\") " pod="cert-manager/cert-manager-86cb77c54b-hq8kc" Dec 03 00:22:24 crc kubenswrapper[4953]: I1203 00:22:24.804105 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rv25m\" (UniqueName: \"kubernetes.io/projected/d6305126-b142-45e0-a058-1356b122cecd-kube-api-access-rv25m\") pod \"cert-manager-86cb77c54b-hq8kc\" (UID: \"d6305126-b142-45e0-a058-1356b122cecd\") " pod="cert-manager/cert-manager-86cb77c54b-hq8kc" Dec 03 00:22:24 crc kubenswrapper[4953]: I1203 00:22:24.825630 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/d6305126-b142-45e0-a058-1356b122cecd-bound-sa-token\") pod \"cert-manager-86cb77c54b-hq8kc\" (UID: \"d6305126-b142-45e0-a058-1356b122cecd\") " pod="cert-manager/cert-manager-86cb77c54b-hq8kc" Dec 03 00:22:24 crc kubenswrapper[4953]: I1203 00:22:24.826633 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rv25m\" (UniqueName: \"kubernetes.io/projected/d6305126-b142-45e0-a058-1356b122cecd-kube-api-access-rv25m\") pod \"cert-manager-86cb77c54b-hq8kc\" (UID: \"d6305126-b142-45e0-a058-1356b122cecd\") " pod="cert-manager/cert-manager-86cb77c54b-hq8kc" Dec 03 00:22:24 crc kubenswrapper[4953]: I1203 00:22:24.891939 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-86cb77c54b-hq8kc" Dec 03 00:22:27 crc kubenswrapper[4953]: I1203 00:22:27.427009 4953 generic.go:334] "Generic (PLEG): container finished" podID="3313dc1f-c9a1-4617-a9d6-cdf983a9d618" containerID="ceffa48b195f4207fcf6d3e94fdd111e3a4a18b49d939c9bb514560b02a6feab" exitCode=0 Dec 03 00:22:27 crc kubenswrapper[4953]: I1203 00:22:27.427094 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/elasticsearch-es-default-0" event={"ID":"3313dc1f-c9a1-4617-a9d6-cdf983a9d618","Type":"ContainerDied","Data":"ceffa48b195f4207fcf6d3e94fdd111e3a4a18b49d939c9bb514560b02a6feab"} Dec 03 00:22:33 crc kubenswrapper[4953]: E1203 00:22:33.606763 4953 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/cert-manager/jetstack-cert-manager-rhel9@sha256:29a0fa1c2f2a6cee62a0468a3883d16d491b4af29130dad6e3e2bb2948f274df" Dec 03 00:22:33 crc kubenswrapper[4953]: E1203 00:22:33.607884 4953 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:cert-manager-webhook,Image:registry.redhat.io/cert-manager/jetstack-cert-manager-rhel9@sha256:29a0fa1c2f2a6cee62a0468a3883d16d491b4af29130dad6e3e2bb2948f274df,Command:[/app/cmd/webhook/webhook],Args:[--dynamic-serving-ca-secret-name=cert-manager-webhook-ca --dynamic-serving-ca-secret-namespace=$(POD_NAMESPACE) --dynamic-serving-dns-names=cert-manager-webhook,cert-manager-webhook.$(POD_NAMESPACE),cert-manager-webhook.$(POD_NAMESPACE).svc --secure-port=10250 --v=2],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:10250,Protocol:TCP,HostIP:,},ContainerPort{Name:healthcheck,HostPort:0,ContainerPort:6080,Protocol:TCP,HostIP:,},ContainerPort{Name:http-metrics,HostPort:0,ContainerPort:9402,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:POD_NAMESPACE,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:bound-sa-token,ReadOnly:true,MountPath:/var/run/secrets/openshift/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-5ph8k,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/livez,Port:{1 0 healthcheck},Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:60,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{1 0 healthcheck},Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:5,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000690000,RunAsNonRoot:nil,ReadOnlyRootFilesystem:*true,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod cert-manager-webhook-f4fb5df64-cg7mx_cert-manager(37a3ced5-065d-4be8-a3e9-54deee67aefa): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 03 00:22:33 crc kubenswrapper[4953]: E1203 00:22:33.609038 4953 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cert-manager-webhook\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="cert-manager/cert-manager-webhook-f4fb5df64-cg7mx" podUID="37a3ced5-065d-4be8-a3e9-54deee67aefa" Dec 03 00:22:34 crc kubenswrapper[4953]: I1203 00:22:34.115931 4953 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-86cb77c54b-hq8kc"] Dec 03 00:22:34 crc kubenswrapper[4953]: W1203 00:22:34.122448 4953 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd6305126_b142_45e0_a058_1356b122cecd.slice/crio-b17768db17e9114072de9962bdbc5b4d18141e0b21a2438aabb0b47c12882e65 WatchSource:0}: Error finding container b17768db17e9114072de9962bdbc5b4d18141e0b21a2438aabb0b47c12882e65: Status 404 returned error can't find the container with id b17768db17e9114072de9962bdbc5b4d18141e0b21a2438aabb0b47c12882e65 Dec 03 00:22:34 crc kubenswrapper[4953]: I1203 00:22:34.665982 4953 generic.go:334] "Generic (PLEG): container finished" podID="3313dc1f-c9a1-4617-a9d6-cdf983a9d618" containerID="000bd77e98352259188050fdb463cc060c622fea7237c7a23fad8a272bd2d54d" exitCode=0 Dec 03 00:22:34 crc kubenswrapper[4953]: I1203 00:22:34.666040 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/elasticsearch-es-default-0" event={"ID":"3313dc1f-c9a1-4617-a9d6-cdf983a9d618","Type":"ContainerDied","Data":"000bd77e98352259188050fdb463cc060c622fea7237c7a23fad8a272bd2d54d"} Dec 03 00:22:34 crc kubenswrapper[4953]: I1203 00:22:34.668565 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-855d9ccff4-gqhrh" event={"ID":"e2c5f086-eaf3-49ae-b67c-e3eb601c2c58","Type":"ContainerStarted","Data":"a584f9e433cc36785a4f483e931f5c6bc634261bd9704b80d3ff6c6950666069"} Dec 03 00:22:34 crc kubenswrapper[4953]: I1203 00:22:34.673718 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-86cb77c54b-hq8kc" event={"ID":"d6305126-b142-45e0-a058-1356b122cecd","Type":"ContainerStarted","Data":"b337a0c08d63ac1559b5ee6add74e326e667b4db3610b58a42ecad356ce16e48"} Dec 03 00:22:34 crc kubenswrapper[4953]: I1203 00:22:34.673748 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-86cb77c54b-hq8kc" event={"ID":"d6305126-b142-45e0-a058-1356b122cecd","Type":"ContainerStarted","Data":"b17768db17e9114072de9962bdbc5b4d18141e0b21a2438aabb0b47c12882e65"} Dec 03 00:22:34 crc kubenswrapper[4953]: I1203 00:22:34.677701 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/service-telemetry-operator-2-build" event={"ID":"6519fd57-6635-41d7-b5e0-2fb21731574f","Type":"ContainerStarted","Data":"ec35487aaaad76cb397c2a32f78a74eec9a50e904e9e8cc07587ce1a75b17fe5"} Dec 03 00:22:34 crc kubenswrapper[4953]: I1203 00:22:34.764340 4953 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="service-telemetry/service-telemetry-operator-2-build" podStartSLOduration=28.345099951999998 podStartE2EDuration="37.764323491s" podCreationTimestamp="2025-12-03 00:21:57 +0000 UTC" firstStartedPulling="2025-12-03 00:22:02.641034184 +0000 UTC m=+911.625314967" lastFinishedPulling="2025-12-03 00:22:12.060257723 +0000 UTC m=+921.044538506" observedRunningTime="2025-12-03 00:22:34.763205143 +0000 UTC m=+943.747485936" watchObservedRunningTime="2025-12-03 00:22:34.764323491 +0000 UTC m=+943.748604274" Dec 03 00:22:34 crc kubenswrapper[4953]: I1203 00:22:34.787368 4953 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-cainjector-855d9ccff4-gqhrh" podStartSLOduration=2.111826469 podStartE2EDuration="14.787349741s" podCreationTimestamp="2025-12-03 00:22:20 +0000 UTC" firstStartedPulling="2025-12-03 00:22:21.054162694 +0000 UTC m=+930.038443477" lastFinishedPulling="2025-12-03 00:22:33.729685956 +0000 UTC m=+942.713966749" observedRunningTime="2025-12-03 00:22:34.784004158 +0000 UTC m=+943.768284941" watchObservedRunningTime="2025-12-03 00:22:34.787349741 +0000 UTC m=+943.771630514" Dec 03 00:22:41 crc kubenswrapper[4953]: I1203 00:22:41.724910 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-f4fb5df64-cg7mx" event={"ID":"37a3ced5-065d-4be8-a3e9-54deee67aefa","Type":"ContainerStarted","Data":"67d51ad7e41e0bdd7a1b3697e2789a2a34492367283ea977287a9ca23f4072cc"} Dec 03 00:22:41 crc kubenswrapper[4953]: I1203 00:22:41.726583 4953 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="cert-manager/cert-manager-webhook-f4fb5df64-cg7mx" Dec 03 00:22:41 crc kubenswrapper[4953]: I1203 00:22:41.733246 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/elasticsearch-es-default-0" event={"ID":"3313dc1f-c9a1-4617-a9d6-cdf983a9d618","Type":"ContainerStarted","Data":"694b505f25c269ba7ec6a39e2a14c231db814aa5331163b286eb6ad9b0e5542c"} Dec 03 00:22:41 crc kubenswrapper[4953]: I1203 00:22:41.733712 4953 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="service-telemetry/elasticsearch-es-default-0" Dec 03 00:22:41 crc kubenswrapper[4953]: I1203 00:22:41.772663 4953 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-webhook-f4fb5df64-cg7mx" podStartSLOduration=-9223372011.082134 podStartE2EDuration="25.772642163s" podCreationTimestamp="2025-12-03 00:22:16 +0000 UTC" firstStartedPulling="2025-12-03 00:22:17.661058942 +0000 UTC m=+926.645339725" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 00:22:41.766004888 +0000 UTC m=+950.750285681" watchObservedRunningTime="2025-12-03 00:22:41.772642163 +0000 UTC m=+950.756922946" Dec 03 00:22:41 crc kubenswrapper[4953]: I1203 00:22:41.870916 4953 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="service-telemetry/elasticsearch-es-default-0" podStartSLOduration=21.489489383 podStartE2EDuration="1m7.870891035s" podCreationTimestamp="2025-12-03 00:21:34 +0000 UTC" firstStartedPulling="2025-12-03 00:21:35.433869315 +0000 UTC m=+884.418150098" lastFinishedPulling="2025-12-03 00:22:21.815270967 +0000 UTC m=+930.799551750" observedRunningTime="2025-12-03 00:22:41.866213689 +0000 UTC m=+950.850494492" watchObservedRunningTime="2025-12-03 00:22:41.870891035 +0000 UTC m=+950.855171818" Dec 03 00:22:41 crc kubenswrapper[4953]: I1203 00:22:41.871164 4953 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-86cb77c54b-hq8kc" podStartSLOduration=17.871157822 podStartE2EDuration="17.871157822s" podCreationTimestamp="2025-12-03 00:22:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 00:22:41.788651119 +0000 UTC m=+950.772931902" watchObservedRunningTime="2025-12-03 00:22:41.871157822 +0000 UTC m=+950.855438925" Dec 03 00:22:46 crc kubenswrapper[4953]: I1203 00:22:46.808467 4953 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="cert-manager/cert-manager-webhook-f4fb5df64-cg7mx" Dec 03 00:22:49 crc kubenswrapper[4953]: I1203 00:22:49.036467 4953 patch_prober.go:28] interesting pod/machine-config-daemon-6dfbn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 00:22:49 crc kubenswrapper[4953]: I1203 00:22:49.037209 4953 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-6dfbn" podUID="53ffb1e9-9a2e-4869-80d7-776f4aa7ef8b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 00:22:55 crc kubenswrapper[4953]: I1203 00:22:55.008431 4953 prober.go:107] "Probe failed" probeType="Readiness" pod="service-telemetry/elasticsearch-es-default-0" podUID="3313dc1f-c9a1-4617-a9d6-cdf983a9d618" containerName="elasticsearch" probeResult="failure" output=< Dec 03 00:22:55 crc kubenswrapper[4953]: {"timestamp": "2025-12-03T00:22:54+00:00", "message": "readiness probe failed", "curl_rc": "7"} Dec 03 00:22:55 crc kubenswrapper[4953]: > Dec 03 00:22:58 crc kubenswrapper[4953]: I1203 00:22:58.866473 4953 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-dpjwn"] Dec 03 00:22:58 crc kubenswrapper[4953]: I1203 00:22:58.868633 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-dpjwn" Dec 03 00:22:58 crc kubenswrapper[4953]: I1203 00:22:58.880086 4953 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-dpjwn"] Dec 03 00:22:58 crc kubenswrapper[4953]: I1203 00:22:58.896914 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c1290040-55a6-4f4c-87f6-dfce258bf189-utilities\") pod \"community-operators-dpjwn\" (UID: \"c1290040-55a6-4f4c-87f6-dfce258bf189\") " pod="openshift-marketplace/community-operators-dpjwn" Dec 03 00:22:58 crc kubenswrapper[4953]: I1203 00:22:58.897034 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c1290040-55a6-4f4c-87f6-dfce258bf189-catalog-content\") pod \"community-operators-dpjwn\" (UID: \"c1290040-55a6-4f4c-87f6-dfce258bf189\") " pod="openshift-marketplace/community-operators-dpjwn" Dec 03 00:22:58 crc kubenswrapper[4953]: I1203 00:22:58.897067 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k6vhj\" (UniqueName: \"kubernetes.io/projected/c1290040-55a6-4f4c-87f6-dfce258bf189-kube-api-access-k6vhj\") pod \"community-operators-dpjwn\" (UID: \"c1290040-55a6-4f4c-87f6-dfce258bf189\") " pod="openshift-marketplace/community-operators-dpjwn" Dec 03 00:22:58 crc kubenswrapper[4953]: I1203 00:22:58.998820 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c1290040-55a6-4f4c-87f6-dfce258bf189-utilities\") pod \"community-operators-dpjwn\" (UID: \"c1290040-55a6-4f4c-87f6-dfce258bf189\") " pod="openshift-marketplace/community-operators-dpjwn" Dec 03 00:22:58 crc kubenswrapper[4953]: I1203 00:22:58.998917 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c1290040-55a6-4f4c-87f6-dfce258bf189-catalog-content\") pod \"community-operators-dpjwn\" (UID: \"c1290040-55a6-4f4c-87f6-dfce258bf189\") " pod="openshift-marketplace/community-operators-dpjwn" Dec 03 00:22:58 crc kubenswrapper[4953]: I1203 00:22:58.998950 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k6vhj\" (UniqueName: \"kubernetes.io/projected/c1290040-55a6-4f4c-87f6-dfce258bf189-kube-api-access-k6vhj\") pod \"community-operators-dpjwn\" (UID: \"c1290040-55a6-4f4c-87f6-dfce258bf189\") " pod="openshift-marketplace/community-operators-dpjwn" Dec 03 00:22:58 crc kubenswrapper[4953]: I1203 00:22:58.999692 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c1290040-55a6-4f4c-87f6-dfce258bf189-utilities\") pod \"community-operators-dpjwn\" (UID: \"c1290040-55a6-4f4c-87f6-dfce258bf189\") " pod="openshift-marketplace/community-operators-dpjwn" Dec 03 00:22:59 crc kubenswrapper[4953]: I1203 00:22:59.000283 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c1290040-55a6-4f4c-87f6-dfce258bf189-catalog-content\") pod \"community-operators-dpjwn\" (UID: \"c1290040-55a6-4f4c-87f6-dfce258bf189\") " pod="openshift-marketplace/community-operators-dpjwn" Dec 03 00:22:59 crc kubenswrapper[4953]: I1203 00:22:59.034990 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k6vhj\" (UniqueName: \"kubernetes.io/projected/c1290040-55a6-4f4c-87f6-dfce258bf189-kube-api-access-k6vhj\") pod \"community-operators-dpjwn\" (UID: \"c1290040-55a6-4f4c-87f6-dfce258bf189\") " pod="openshift-marketplace/community-operators-dpjwn" Dec 03 00:22:59 crc kubenswrapper[4953]: I1203 00:22:59.199881 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-dpjwn" Dec 03 00:22:59 crc kubenswrapper[4953]: I1203 00:22:59.987428 4953 prober.go:107] "Probe failed" probeType="Readiness" pod="service-telemetry/elasticsearch-es-default-0" podUID="3313dc1f-c9a1-4617-a9d6-cdf983a9d618" containerName="elasticsearch" probeResult="failure" output=< Dec 03 00:22:59 crc kubenswrapper[4953]: {"timestamp": "2025-12-03T00:22:59+00:00", "message": "readiness probe failed", "curl_rc": "7"} Dec 03 00:22:59 crc kubenswrapper[4953]: > Dec 03 00:23:00 crc kubenswrapper[4953]: I1203 00:23:00.058759 4953 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-dpjwn"] Dec 03 00:23:01 crc kubenswrapper[4953]: I1203 00:23:01.080475 4953 generic.go:334] "Generic (PLEG): container finished" podID="c1290040-55a6-4f4c-87f6-dfce258bf189" containerID="48e84c0d0be8d20647afde29186ed7f0194e4029a8e05faa1a7a3915fb0f1ce5" exitCode=0 Dec 03 00:23:01 crc kubenswrapper[4953]: I1203 00:23:01.080559 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-dpjwn" event={"ID":"c1290040-55a6-4f4c-87f6-dfce258bf189","Type":"ContainerDied","Data":"48e84c0d0be8d20647afde29186ed7f0194e4029a8e05faa1a7a3915fb0f1ce5"} Dec 03 00:23:01 crc kubenswrapper[4953]: I1203 00:23:01.080596 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-dpjwn" event={"ID":"c1290040-55a6-4f4c-87f6-dfce258bf189","Type":"ContainerStarted","Data":"47b3df79998f0d8891597194e8130c32b5435d6602878abaf286579942db1cc0"} Dec 03 00:23:02 crc kubenswrapper[4953]: I1203 00:23:02.090286 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-dpjwn" event={"ID":"c1290040-55a6-4f4c-87f6-dfce258bf189","Type":"ContainerStarted","Data":"d357f20eaa0aa74fb49763193bc0f730025fbf50c52343f40b4774af80fa6b8d"} Dec 03 00:23:04 crc kubenswrapper[4953]: I1203 00:23:04.105575 4953 generic.go:334] "Generic (PLEG): container finished" podID="c1290040-55a6-4f4c-87f6-dfce258bf189" containerID="d357f20eaa0aa74fb49763193bc0f730025fbf50c52343f40b4774af80fa6b8d" exitCode=0 Dec 03 00:23:04 crc kubenswrapper[4953]: I1203 00:23:04.105660 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-dpjwn" event={"ID":"c1290040-55a6-4f4c-87f6-dfce258bf189","Type":"ContainerDied","Data":"d357f20eaa0aa74fb49763193bc0f730025fbf50c52343f40b4774af80fa6b8d"} Dec 03 00:23:05 crc kubenswrapper[4953]: I1203 00:23:05.018922 4953 prober.go:107] "Probe failed" probeType="Readiness" pod="service-telemetry/elasticsearch-es-default-0" podUID="3313dc1f-c9a1-4617-a9d6-cdf983a9d618" containerName="elasticsearch" probeResult="failure" output=< Dec 03 00:23:05 crc kubenswrapper[4953]: {"timestamp": "2025-12-03T00:23:05+00:00", "message": "readiness probe failed", "curl_rc": "7"} Dec 03 00:23:05 crc kubenswrapper[4953]: > Dec 03 00:23:05 crc kubenswrapper[4953]: I1203 00:23:05.116987 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-dpjwn" event={"ID":"c1290040-55a6-4f4c-87f6-dfce258bf189","Type":"ContainerStarted","Data":"5c4626439d24f8404e44d82387b51493520c18eb491dda4b99e5b639a9456788"} Dec 03 00:23:05 crc kubenswrapper[4953]: I1203 00:23:05.143458 4953 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-dpjwn" podStartSLOduration=3.538716926 podStartE2EDuration="7.143440707s" podCreationTimestamp="2025-12-03 00:22:58 +0000 UTC" firstStartedPulling="2025-12-03 00:23:01.082473403 +0000 UTC m=+970.066754186" lastFinishedPulling="2025-12-03 00:23:04.687197184 +0000 UTC m=+973.671477967" observedRunningTime="2025-12-03 00:23:05.141691383 +0000 UTC m=+974.125972166" watchObservedRunningTime="2025-12-03 00:23:05.143440707 +0000 UTC m=+974.127721490" Dec 03 00:23:09 crc kubenswrapper[4953]: I1203 00:23:09.200058 4953 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-dpjwn" Dec 03 00:23:09 crc kubenswrapper[4953]: I1203 00:23:09.200141 4953 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-dpjwn" Dec 03 00:23:09 crc kubenswrapper[4953]: I1203 00:23:09.253059 4953 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-dpjwn" Dec 03 00:23:10 crc kubenswrapper[4953]: I1203 00:23:10.263359 4953 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-dpjwn" Dec 03 00:23:10 crc kubenswrapper[4953]: I1203 00:23:10.502127 4953 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="service-telemetry/elasticsearch-es-default-0" Dec 03 00:23:10 crc kubenswrapper[4953]: I1203 00:23:10.642289 4953 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-dpjwn"] Dec 03 00:23:12 crc kubenswrapper[4953]: I1203 00:23:12.222353 4953 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-dpjwn" podUID="c1290040-55a6-4f4c-87f6-dfce258bf189" containerName="registry-server" containerID="cri-o://5c4626439d24f8404e44d82387b51493520c18eb491dda4b99e5b639a9456788" gracePeriod=2 Dec 03 00:23:13 crc kubenswrapper[4953]: I1203 00:23:13.035027 4953 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-dpjwn" Dec 03 00:23:13 crc kubenswrapper[4953]: I1203 00:23:13.098841 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c1290040-55a6-4f4c-87f6-dfce258bf189-utilities\") pod \"c1290040-55a6-4f4c-87f6-dfce258bf189\" (UID: \"c1290040-55a6-4f4c-87f6-dfce258bf189\") " Dec 03 00:23:13 crc kubenswrapper[4953]: I1203 00:23:13.098958 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c1290040-55a6-4f4c-87f6-dfce258bf189-catalog-content\") pod \"c1290040-55a6-4f4c-87f6-dfce258bf189\" (UID: \"c1290040-55a6-4f4c-87f6-dfce258bf189\") " Dec 03 00:23:13 crc kubenswrapper[4953]: I1203 00:23:13.099029 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k6vhj\" (UniqueName: \"kubernetes.io/projected/c1290040-55a6-4f4c-87f6-dfce258bf189-kube-api-access-k6vhj\") pod \"c1290040-55a6-4f4c-87f6-dfce258bf189\" (UID: \"c1290040-55a6-4f4c-87f6-dfce258bf189\") " Dec 03 00:23:13 crc kubenswrapper[4953]: I1203 00:23:13.116315 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c1290040-55a6-4f4c-87f6-dfce258bf189-utilities" (OuterVolumeSpecName: "utilities") pod "c1290040-55a6-4f4c-87f6-dfce258bf189" (UID: "c1290040-55a6-4f4c-87f6-dfce258bf189"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 00:23:13 crc kubenswrapper[4953]: I1203 00:23:13.126041 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c1290040-55a6-4f4c-87f6-dfce258bf189-kube-api-access-k6vhj" (OuterVolumeSpecName: "kube-api-access-k6vhj") pod "c1290040-55a6-4f4c-87f6-dfce258bf189" (UID: "c1290040-55a6-4f4c-87f6-dfce258bf189"). InnerVolumeSpecName "kube-api-access-k6vhj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:23:13 crc kubenswrapper[4953]: I1203 00:23:13.168962 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c1290040-55a6-4f4c-87f6-dfce258bf189-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c1290040-55a6-4f4c-87f6-dfce258bf189" (UID: "c1290040-55a6-4f4c-87f6-dfce258bf189"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 00:23:13 crc kubenswrapper[4953]: I1203 00:23:13.201511 4953 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c1290040-55a6-4f4c-87f6-dfce258bf189-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 00:23:13 crc kubenswrapper[4953]: I1203 00:23:13.201562 4953 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c1290040-55a6-4f4c-87f6-dfce258bf189-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 00:23:13 crc kubenswrapper[4953]: I1203 00:23:13.201579 4953 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k6vhj\" (UniqueName: \"kubernetes.io/projected/c1290040-55a6-4f4c-87f6-dfce258bf189-kube-api-access-k6vhj\") on node \"crc\" DevicePath \"\"" Dec 03 00:23:13 crc kubenswrapper[4953]: I1203 00:23:13.233284 4953 generic.go:334] "Generic (PLEG): container finished" podID="c1290040-55a6-4f4c-87f6-dfce258bf189" containerID="5c4626439d24f8404e44d82387b51493520c18eb491dda4b99e5b639a9456788" exitCode=0 Dec 03 00:23:13 crc kubenswrapper[4953]: I1203 00:23:13.233355 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-dpjwn" event={"ID":"c1290040-55a6-4f4c-87f6-dfce258bf189","Type":"ContainerDied","Data":"5c4626439d24f8404e44d82387b51493520c18eb491dda4b99e5b639a9456788"} Dec 03 00:23:13 crc kubenswrapper[4953]: I1203 00:23:13.233399 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-dpjwn" event={"ID":"c1290040-55a6-4f4c-87f6-dfce258bf189","Type":"ContainerDied","Data":"47b3df79998f0d8891597194e8130c32b5435d6602878abaf286579942db1cc0"} Dec 03 00:23:13 crc kubenswrapper[4953]: I1203 00:23:13.233426 4953 scope.go:117] "RemoveContainer" containerID="5c4626439d24f8404e44d82387b51493520c18eb491dda4b99e5b639a9456788" Dec 03 00:23:13 crc kubenswrapper[4953]: I1203 00:23:13.233364 4953 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-dpjwn" Dec 03 00:23:13 crc kubenswrapper[4953]: I1203 00:23:13.254565 4953 scope.go:117] "RemoveContainer" containerID="d357f20eaa0aa74fb49763193bc0f730025fbf50c52343f40b4774af80fa6b8d" Dec 03 00:23:13 crc kubenswrapper[4953]: I1203 00:23:13.284491 4953 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-dpjwn"] Dec 03 00:23:13 crc kubenswrapper[4953]: I1203 00:23:13.299740 4953 scope.go:117] "RemoveContainer" containerID="48e84c0d0be8d20647afde29186ed7f0194e4029a8e05faa1a7a3915fb0f1ce5" Dec 03 00:23:13 crc kubenswrapper[4953]: I1203 00:23:13.300311 4953 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-dpjwn"] Dec 03 00:23:13 crc kubenswrapper[4953]: I1203 00:23:13.324801 4953 scope.go:117] "RemoveContainer" containerID="5c4626439d24f8404e44d82387b51493520c18eb491dda4b99e5b639a9456788" Dec 03 00:23:13 crc kubenswrapper[4953]: E1203 00:23:13.325537 4953 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5c4626439d24f8404e44d82387b51493520c18eb491dda4b99e5b639a9456788\": container with ID starting with 5c4626439d24f8404e44d82387b51493520c18eb491dda4b99e5b639a9456788 not found: ID does not exist" containerID="5c4626439d24f8404e44d82387b51493520c18eb491dda4b99e5b639a9456788" Dec 03 00:23:13 crc kubenswrapper[4953]: I1203 00:23:13.325584 4953 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5c4626439d24f8404e44d82387b51493520c18eb491dda4b99e5b639a9456788"} err="failed to get container status \"5c4626439d24f8404e44d82387b51493520c18eb491dda4b99e5b639a9456788\": rpc error: code = NotFound desc = could not find container \"5c4626439d24f8404e44d82387b51493520c18eb491dda4b99e5b639a9456788\": container with ID starting with 5c4626439d24f8404e44d82387b51493520c18eb491dda4b99e5b639a9456788 not found: ID does not exist" Dec 03 00:23:13 crc kubenswrapper[4953]: I1203 00:23:13.325620 4953 scope.go:117] "RemoveContainer" containerID="d357f20eaa0aa74fb49763193bc0f730025fbf50c52343f40b4774af80fa6b8d" Dec 03 00:23:13 crc kubenswrapper[4953]: E1203 00:23:13.325904 4953 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d357f20eaa0aa74fb49763193bc0f730025fbf50c52343f40b4774af80fa6b8d\": container with ID starting with d357f20eaa0aa74fb49763193bc0f730025fbf50c52343f40b4774af80fa6b8d not found: ID does not exist" containerID="d357f20eaa0aa74fb49763193bc0f730025fbf50c52343f40b4774af80fa6b8d" Dec 03 00:23:13 crc kubenswrapper[4953]: I1203 00:23:13.325959 4953 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d357f20eaa0aa74fb49763193bc0f730025fbf50c52343f40b4774af80fa6b8d"} err="failed to get container status \"d357f20eaa0aa74fb49763193bc0f730025fbf50c52343f40b4774af80fa6b8d\": rpc error: code = NotFound desc = could not find container \"d357f20eaa0aa74fb49763193bc0f730025fbf50c52343f40b4774af80fa6b8d\": container with ID starting with d357f20eaa0aa74fb49763193bc0f730025fbf50c52343f40b4774af80fa6b8d not found: ID does not exist" Dec 03 00:23:13 crc kubenswrapper[4953]: I1203 00:23:13.326016 4953 scope.go:117] "RemoveContainer" containerID="48e84c0d0be8d20647afde29186ed7f0194e4029a8e05faa1a7a3915fb0f1ce5" Dec 03 00:23:13 crc kubenswrapper[4953]: E1203 00:23:13.327600 4953 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"48e84c0d0be8d20647afde29186ed7f0194e4029a8e05faa1a7a3915fb0f1ce5\": container with ID starting with 48e84c0d0be8d20647afde29186ed7f0194e4029a8e05faa1a7a3915fb0f1ce5 not found: ID does not exist" containerID="48e84c0d0be8d20647afde29186ed7f0194e4029a8e05faa1a7a3915fb0f1ce5" Dec 03 00:23:13 crc kubenswrapper[4953]: I1203 00:23:13.327632 4953 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"48e84c0d0be8d20647afde29186ed7f0194e4029a8e05faa1a7a3915fb0f1ce5"} err="failed to get container status \"48e84c0d0be8d20647afde29186ed7f0194e4029a8e05faa1a7a3915fb0f1ce5\": rpc error: code = NotFound desc = could not find container \"48e84c0d0be8d20647afde29186ed7f0194e4029a8e05faa1a7a3915fb0f1ce5\": container with ID starting with 48e84c0d0be8d20647afde29186ed7f0194e4029a8e05faa1a7a3915fb0f1ce5 not found: ID does not exist" Dec 03 00:23:13 crc kubenswrapper[4953]: I1203 00:23:13.358942 4953 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c1290040-55a6-4f4c-87f6-dfce258bf189" path="/var/lib/kubelet/pods/c1290040-55a6-4f4c-87f6-dfce258bf189/volumes" Dec 03 00:23:18 crc kubenswrapper[4953]: I1203 00:23:18.944461 4953 patch_prober.go:28] interesting pod/machine-config-daemon-6dfbn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 00:23:18 crc kubenswrapper[4953]: I1203 00:23:18.945405 4953 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-6dfbn" podUID="53ffb1e9-9a2e-4869-80d7-776f4aa7ef8b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 00:23:48 crc kubenswrapper[4953]: I1203 00:23:48.945248 4953 patch_prober.go:28] interesting pod/machine-config-daemon-6dfbn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 00:23:48 crc kubenswrapper[4953]: I1203 00:23:48.946045 4953 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-6dfbn" podUID="53ffb1e9-9a2e-4869-80d7-776f4aa7ef8b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 00:23:48 crc kubenswrapper[4953]: I1203 00:23:48.946112 4953 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-6dfbn" Dec 03 00:23:48 crc kubenswrapper[4953]: I1203 00:23:48.946942 4953 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"ede43397367a8a468be56d98b9f73d5afe6a5294be1471b33f69b64147ab3995"} pod="openshift-machine-config-operator/machine-config-daemon-6dfbn" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 03 00:23:48 crc kubenswrapper[4953]: I1203 00:23:48.947029 4953 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-6dfbn" podUID="53ffb1e9-9a2e-4869-80d7-776f4aa7ef8b" containerName="machine-config-daemon" containerID="cri-o://ede43397367a8a468be56d98b9f73d5afe6a5294be1471b33f69b64147ab3995" gracePeriod=600 Dec 03 00:23:49 crc kubenswrapper[4953]: I1203 00:23:49.603522 4953 generic.go:334] "Generic (PLEG): container finished" podID="53ffb1e9-9a2e-4869-80d7-776f4aa7ef8b" containerID="ede43397367a8a468be56d98b9f73d5afe6a5294be1471b33f69b64147ab3995" exitCode=0 Dec 03 00:23:49 crc kubenswrapper[4953]: I1203 00:23:49.603560 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-6dfbn" event={"ID":"53ffb1e9-9a2e-4869-80d7-776f4aa7ef8b","Type":"ContainerDied","Data":"ede43397367a8a468be56d98b9f73d5afe6a5294be1471b33f69b64147ab3995"} Dec 03 00:23:49 crc kubenswrapper[4953]: I1203 00:23:49.604086 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-6dfbn" event={"ID":"53ffb1e9-9a2e-4869-80d7-776f4aa7ef8b","Type":"ContainerStarted","Data":"1ed646ac60a808beb2f58a1886a69c6cb3f62ef93231fc4501030f895ee3958c"} Dec 03 00:23:49 crc kubenswrapper[4953]: I1203 00:23:49.604121 4953 scope.go:117] "RemoveContainer" containerID="553caefdc3000cd9d8eaf08bf4d23cdb2a37135c2cd1053fce9b7f6cb2b25c0b" Dec 03 00:24:51 crc kubenswrapper[4953]: I1203 00:24:51.046728 4953 generic.go:334] "Generic (PLEG): container finished" podID="6519fd57-6635-41d7-b5e0-2fb21731574f" containerID="ec35487aaaad76cb397c2a32f78a74eec9a50e904e9e8cc07587ce1a75b17fe5" exitCode=0 Dec 03 00:24:51 crc kubenswrapper[4953]: I1203 00:24:51.046836 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/service-telemetry-operator-2-build" event={"ID":"6519fd57-6635-41d7-b5e0-2fb21731574f","Type":"ContainerDied","Data":"ec35487aaaad76cb397c2a32f78a74eec9a50e904e9e8cc07587ce1a75b17fe5"} Dec 03 00:24:52 crc kubenswrapper[4953]: I1203 00:24:52.331877 4953 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/service-telemetry-operator-2-build" Dec 03 00:24:52 crc kubenswrapper[4953]: I1203 00:24:52.417915 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/6519fd57-6635-41d7-b5e0-2fb21731574f-build-system-configs\") pod \"6519fd57-6635-41d7-b5e0-2fb21731574f\" (UID: \"6519fd57-6635-41d7-b5e0-2fb21731574f\") " Dec 03 00:24:52 crc kubenswrapper[4953]: I1203 00:24:52.419082 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/6519fd57-6635-41d7-b5e0-2fb21731574f-node-pullsecrets\") pod \"6519fd57-6635-41d7-b5e0-2fb21731574f\" (UID: \"6519fd57-6635-41d7-b5e0-2fb21731574f\") " Dec 03 00:24:52 crc kubenswrapper[4953]: I1203 00:24:52.419022 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6519fd57-6635-41d7-b5e0-2fb21731574f-build-system-configs" (OuterVolumeSpecName: "build-system-configs") pod "6519fd57-6635-41d7-b5e0-2fb21731574f" (UID: "6519fd57-6635-41d7-b5e0-2fb21731574f"). InnerVolumeSpecName "build-system-configs". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:24:52 crc kubenswrapper[4953]: I1203 00:24:52.419166 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"builder-dockercfg-9prwd-pull\" (UniqueName: \"kubernetes.io/secret/6519fd57-6635-41d7-b5e0-2fb21731574f-builder-dockercfg-9prwd-pull\") pod \"6519fd57-6635-41d7-b5e0-2fb21731574f\" (UID: \"6519fd57-6635-41d7-b5e0-2fb21731574f\") " Dec 03 00:24:52 crc kubenswrapper[4953]: I1203 00:24:52.419213 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/6519fd57-6635-41d7-b5e0-2fb21731574f-node-pullsecrets" (OuterVolumeSpecName: "node-pullsecrets") pod "6519fd57-6635-41d7-b5e0-2fb21731574f" (UID: "6519fd57-6635-41d7-b5e0-2fb21731574f"). InnerVolumeSpecName "node-pullsecrets". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 00:24:52 crc kubenswrapper[4953]: I1203 00:24:52.419233 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/6519fd57-6635-41d7-b5e0-2fb21731574f-buildworkdir\") pod \"6519fd57-6635-41d7-b5e0-2fb21731574f\" (UID: \"6519fd57-6635-41d7-b5e0-2fb21731574f\") " Dec 03 00:24:52 crc kubenswrapper[4953]: I1203 00:24:52.419253 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"builder-dockercfg-9prwd-push\" (UniqueName: \"kubernetes.io/secret/6519fd57-6635-41d7-b5e0-2fb21731574f-builder-dockercfg-9prwd-push\") pod \"6519fd57-6635-41d7-b5e0-2fb21731574f\" (UID: \"6519fd57-6635-41d7-b5e0-2fb21731574f\") " Dec 03 00:24:52 crc kubenswrapper[4953]: I1203 00:24:52.420612 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6519fd57-6635-41d7-b5e0-2fb21731574f-container-storage-run" (OuterVolumeSpecName: "container-storage-run") pod "6519fd57-6635-41d7-b5e0-2fb21731574f" (UID: "6519fd57-6635-41d7-b5e0-2fb21731574f"). InnerVolumeSpecName "container-storage-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 00:24:52 crc kubenswrapper[4953]: I1203 00:24:52.419276 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/6519fd57-6635-41d7-b5e0-2fb21731574f-container-storage-run\") pod \"6519fd57-6635-41d7-b5e0-2fb21731574f\" (UID: \"6519fd57-6635-41d7-b5e0-2fb21731574f\") " Dec 03 00:24:52 crc kubenswrapper[4953]: I1203 00:24:52.420944 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4ljpw\" (UniqueName: \"kubernetes.io/projected/6519fd57-6635-41d7-b5e0-2fb21731574f-kube-api-access-4ljpw\") pod \"6519fd57-6635-41d7-b5e0-2fb21731574f\" (UID: \"6519fd57-6635-41d7-b5e0-2fb21731574f\") " Dec 03 00:24:52 crc kubenswrapper[4953]: I1203 00:24:52.421089 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/6519fd57-6635-41d7-b5e0-2fb21731574f-container-storage-root\") pod \"6519fd57-6635-41d7-b5e0-2fb21731574f\" (UID: \"6519fd57-6635-41d7-b5e0-2fb21731574f\") " Dec 03 00:24:52 crc kubenswrapper[4953]: I1203 00:24:52.421154 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/6519fd57-6635-41d7-b5e0-2fb21731574f-buildcachedir\") pod \"6519fd57-6635-41d7-b5e0-2fb21731574f\" (UID: \"6519fd57-6635-41d7-b5e0-2fb21731574f\") " Dec 03 00:24:52 crc kubenswrapper[4953]: I1203 00:24:52.421223 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/6519fd57-6635-41d7-b5e0-2fb21731574f-buildcachedir" (OuterVolumeSpecName: "buildcachedir") pod "6519fd57-6635-41d7-b5e0-2fb21731574f" (UID: "6519fd57-6635-41d7-b5e0-2fb21731574f"). InnerVolumeSpecName "buildcachedir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 00:24:52 crc kubenswrapper[4953]: I1203 00:24:52.421302 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/6519fd57-6635-41d7-b5e0-2fb21731574f-build-ca-bundles\") pod \"6519fd57-6635-41d7-b5e0-2fb21731574f\" (UID: \"6519fd57-6635-41d7-b5e0-2fb21731574f\") " Dec 03 00:24:52 crc kubenswrapper[4953]: I1203 00:24:52.421914 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6519fd57-6635-41d7-b5e0-2fb21731574f-build-ca-bundles" (OuterVolumeSpecName: "build-ca-bundles") pod "6519fd57-6635-41d7-b5e0-2fb21731574f" (UID: "6519fd57-6635-41d7-b5e0-2fb21731574f"). InnerVolumeSpecName "build-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:24:52 crc kubenswrapper[4953]: I1203 00:24:52.426458 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6519fd57-6635-41d7-b5e0-2fb21731574f-kube-api-access-4ljpw" (OuterVolumeSpecName: "kube-api-access-4ljpw") pod "6519fd57-6635-41d7-b5e0-2fb21731574f" (UID: "6519fd57-6635-41d7-b5e0-2fb21731574f"). InnerVolumeSpecName "kube-api-access-4ljpw". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:24:52 crc kubenswrapper[4953]: I1203 00:24:52.429192 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6519fd57-6635-41d7-b5e0-2fb21731574f-builder-dockercfg-9prwd-push" (OuterVolumeSpecName: "builder-dockercfg-9prwd-push") pod "6519fd57-6635-41d7-b5e0-2fb21731574f" (UID: "6519fd57-6635-41d7-b5e0-2fb21731574f"). InnerVolumeSpecName "builder-dockercfg-9prwd-push". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:24:52 crc kubenswrapper[4953]: I1203 00:24:52.429321 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/6519fd57-6635-41d7-b5e0-2fb21731574f-build-blob-cache\") pod \"6519fd57-6635-41d7-b5e0-2fb21731574f\" (UID: \"6519fd57-6635-41d7-b5e0-2fb21731574f\") " Dec 03 00:24:52 crc kubenswrapper[4953]: I1203 00:24:52.429399 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/6519fd57-6635-41d7-b5e0-2fb21731574f-build-proxy-ca-bundles\") pod \"6519fd57-6635-41d7-b5e0-2fb21731574f\" (UID: \"6519fd57-6635-41d7-b5e0-2fb21731574f\") " Dec 03 00:24:52 crc kubenswrapper[4953]: I1203 00:24:52.430081 4953 reconciler_common.go:293] "Volume detached for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/6519fd57-6635-41d7-b5e0-2fb21731574f-buildcachedir\") on node \"crc\" DevicePath \"\"" Dec 03 00:24:52 crc kubenswrapper[4953]: I1203 00:24:52.430110 4953 reconciler_common.go:293] "Volume detached for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/6519fd57-6635-41d7-b5e0-2fb21731574f-build-system-configs\") on node \"crc\" DevicePath \"\"" Dec 03 00:24:52 crc kubenswrapper[4953]: I1203 00:24:52.430126 4953 reconciler_common.go:293] "Volume detached for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/6519fd57-6635-41d7-b5e0-2fb21731574f-node-pullsecrets\") on node \"crc\" DevicePath \"\"" Dec 03 00:24:52 crc kubenswrapper[4953]: I1203 00:24:52.430145 4953 reconciler_common.go:293] "Volume detached for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/6519fd57-6635-41d7-b5e0-2fb21731574f-container-storage-run\") on node \"crc\" DevicePath \"\"" Dec 03 00:24:52 crc kubenswrapper[4953]: I1203 00:24:52.431196 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6519fd57-6635-41d7-b5e0-2fb21731574f-builder-dockercfg-9prwd-pull" (OuterVolumeSpecName: "builder-dockercfg-9prwd-pull") pod "6519fd57-6635-41d7-b5e0-2fb21731574f" (UID: "6519fd57-6635-41d7-b5e0-2fb21731574f"). InnerVolumeSpecName "builder-dockercfg-9prwd-pull". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:24:52 crc kubenswrapper[4953]: I1203 00:24:52.431368 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6519fd57-6635-41d7-b5e0-2fb21731574f-build-proxy-ca-bundles" (OuterVolumeSpecName: "build-proxy-ca-bundles") pod "6519fd57-6635-41d7-b5e0-2fb21731574f" (UID: "6519fd57-6635-41d7-b5e0-2fb21731574f"). InnerVolumeSpecName "build-proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:24:52 crc kubenswrapper[4953]: I1203 00:24:52.458363 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6519fd57-6635-41d7-b5e0-2fb21731574f-buildworkdir" (OuterVolumeSpecName: "buildworkdir") pod "6519fd57-6635-41d7-b5e0-2fb21731574f" (UID: "6519fd57-6635-41d7-b5e0-2fb21731574f"). InnerVolumeSpecName "buildworkdir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 00:24:52 crc kubenswrapper[4953]: I1203 00:24:52.531567 4953 reconciler_common.go:293] "Volume detached for volume \"builder-dockercfg-9prwd-pull\" (UniqueName: \"kubernetes.io/secret/6519fd57-6635-41d7-b5e0-2fb21731574f-builder-dockercfg-9prwd-pull\") on node \"crc\" DevicePath \"\"" Dec 03 00:24:52 crc kubenswrapper[4953]: I1203 00:24:52.531893 4953 reconciler_common.go:293] "Volume detached for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/6519fd57-6635-41d7-b5e0-2fb21731574f-buildworkdir\") on node \"crc\" DevicePath \"\"" Dec 03 00:24:52 crc kubenswrapper[4953]: I1203 00:24:52.532045 4953 reconciler_common.go:293] "Volume detached for volume \"builder-dockercfg-9prwd-push\" (UniqueName: \"kubernetes.io/secret/6519fd57-6635-41d7-b5e0-2fb21731574f-builder-dockercfg-9prwd-push\") on node \"crc\" DevicePath \"\"" Dec 03 00:24:52 crc kubenswrapper[4953]: I1203 00:24:52.532142 4953 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4ljpw\" (UniqueName: \"kubernetes.io/projected/6519fd57-6635-41d7-b5e0-2fb21731574f-kube-api-access-4ljpw\") on node \"crc\" DevicePath \"\"" Dec 03 00:24:52 crc kubenswrapper[4953]: I1203 00:24:52.532227 4953 reconciler_common.go:293] "Volume detached for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/6519fd57-6635-41d7-b5e0-2fb21731574f-build-ca-bundles\") on node \"crc\" DevicePath \"\"" Dec 03 00:24:52 crc kubenswrapper[4953]: I1203 00:24:52.532304 4953 reconciler_common.go:293] "Volume detached for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/6519fd57-6635-41d7-b5e0-2fb21731574f-build-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Dec 03 00:24:52 crc kubenswrapper[4953]: I1203 00:24:52.612809 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6519fd57-6635-41d7-b5e0-2fb21731574f-build-blob-cache" (OuterVolumeSpecName: "build-blob-cache") pod "6519fd57-6635-41d7-b5e0-2fb21731574f" (UID: "6519fd57-6635-41d7-b5e0-2fb21731574f"). InnerVolumeSpecName "build-blob-cache". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 00:24:52 crc kubenswrapper[4953]: I1203 00:24:52.634001 4953 reconciler_common.go:293] "Volume detached for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/6519fd57-6635-41d7-b5e0-2fb21731574f-build-blob-cache\") on node \"crc\" DevicePath \"\"" Dec 03 00:24:53 crc kubenswrapper[4953]: I1203 00:24:53.065387 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/service-telemetry-operator-2-build" event={"ID":"6519fd57-6635-41d7-b5e0-2fb21731574f","Type":"ContainerDied","Data":"07c4a3a7f333bcf6a1530d2246b9cffd33d5bcae3f9c90803773c00e1c540078"} Dec 03 00:24:53 crc kubenswrapper[4953]: I1203 00:24:53.065439 4953 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="07c4a3a7f333bcf6a1530d2246b9cffd33d5bcae3f9c90803773c00e1c540078" Dec 03 00:24:53 crc kubenswrapper[4953]: I1203 00:24:53.065509 4953 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/service-telemetry-operator-2-build" Dec 03 00:24:54 crc kubenswrapper[4953]: I1203 00:24:54.785915 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6519fd57-6635-41d7-b5e0-2fb21731574f-container-storage-root" (OuterVolumeSpecName: "container-storage-root") pod "6519fd57-6635-41d7-b5e0-2fb21731574f" (UID: "6519fd57-6635-41d7-b5e0-2fb21731574f"). InnerVolumeSpecName "container-storage-root". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 00:24:54 crc kubenswrapper[4953]: I1203 00:24:54.876948 4953 reconciler_common.go:293] "Volume detached for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/6519fd57-6635-41d7-b5e0-2fb21731574f-container-storage-root\") on node \"crc\" DevicePath \"\"" Dec 03 00:24:57 crc kubenswrapper[4953]: I1203 00:24:57.344214 4953 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["service-telemetry/smart-gateway-operator-1-build"] Dec 03 00:24:57 crc kubenswrapper[4953]: E1203 00:24:57.345076 4953 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c1290040-55a6-4f4c-87f6-dfce258bf189" containerName="registry-server" Dec 03 00:24:57 crc kubenswrapper[4953]: I1203 00:24:57.345100 4953 state_mem.go:107] "Deleted CPUSet assignment" podUID="c1290040-55a6-4f4c-87f6-dfce258bf189" containerName="registry-server" Dec 03 00:24:57 crc kubenswrapper[4953]: E1203 00:24:57.345124 4953 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6519fd57-6635-41d7-b5e0-2fb21731574f" containerName="docker-build" Dec 03 00:24:57 crc kubenswrapper[4953]: I1203 00:24:57.345135 4953 state_mem.go:107] "Deleted CPUSet assignment" podUID="6519fd57-6635-41d7-b5e0-2fb21731574f" containerName="docker-build" Dec 03 00:24:57 crc kubenswrapper[4953]: E1203 00:24:57.345150 4953 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6519fd57-6635-41d7-b5e0-2fb21731574f" containerName="manage-dockerfile" Dec 03 00:24:57 crc kubenswrapper[4953]: I1203 00:24:57.345162 4953 state_mem.go:107] "Deleted CPUSet assignment" podUID="6519fd57-6635-41d7-b5e0-2fb21731574f" containerName="manage-dockerfile" Dec 03 00:24:57 crc kubenswrapper[4953]: E1203 00:24:57.345183 4953 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6519fd57-6635-41d7-b5e0-2fb21731574f" containerName="git-clone" Dec 03 00:24:57 crc kubenswrapper[4953]: I1203 00:24:57.345194 4953 state_mem.go:107] "Deleted CPUSet assignment" podUID="6519fd57-6635-41d7-b5e0-2fb21731574f" containerName="git-clone" Dec 03 00:24:57 crc kubenswrapper[4953]: E1203 00:24:57.345205 4953 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c1290040-55a6-4f4c-87f6-dfce258bf189" containerName="extract-content" Dec 03 00:24:57 crc kubenswrapper[4953]: I1203 00:24:57.345213 4953 state_mem.go:107] "Deleted CPUSet assignment" podUID="c1290040-55a6-4f4c-87f6-dfce258bf189" containerName="extract-content" Dec 03 00:24:57 crc kubenswrapper[4953]: E1203 00:24:57.345232 4953 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c1290040-55a6-4f4c-87f6-dfce258bf189" containerName="extract-utilities" Dec 03 00:24:57 crc kubenswrapper[4953]: I1203 00:24:57.345242 4953 state_mem.go:107] "Deleted CPUSet assignment" podUID="c1290040-55a6-4f4c-87f6-dfce258bf189" containerName="extract-utilities" Dec 03 00:24:57 crc kubenswrapper[4953]: I1203 00:24:57.345461 4953 memory_manager.go:354] "RemoveStaleState removing state" podUID="c1290040-55a6-4f4c-87f6-dfce258bf189" containerName="registry-server" Dec 03 00:24:57 crc kubenswrapper[4953]: I1203 00:24:57.345487 4953 memory_manager.go:354] "RemoveStaleState removing state" podUID="6519fd57-6635-41d7-b5e0-2fb21731574f" containerName="docker-build" Dec 03 00:24:57 crc kubenswrapper[4953]: I1203 00:24:57.346462 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/smart-gateway-operator-1-build" Dec 03 00:24:57 crc kubenswrapper[4953]: I1203 00:24:57.349464 4953 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"builder-dockercfg-9prwd" Dec 03 00:24:57 crc kubenswrapper[4953]: I1203 00:24:57.350116 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"smart-gateway-operator-1-global-ca" Dec 03 00:24:57 crc kubenswrapper[4953]: I1203 00:24:57.350246 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"smart-gateway-operator-1-sys-config" Dec 03 00:24:57 crc kubenswrapper[4953]: I1203 00:24:57.350605 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"smart-gateway-operator-1-ca" Dec 03 00:24:57 crc kubenswrapper[4953]: I1203 00:24:57.360374 4953 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/smart-gateway-operator-1-build"] Dec 03 00:24:57 crc kubenswrapper[4953]: I1203 00:24:57.414480 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"builder-dockercfg-9prwd-pull\" (UniqueName: \"kubernetes.io/secret/2f5ad768-60c4-430b-85c1-90a0f8cddc62-builder-dockercfg-9prwd-pull\") pod \"smart-gateway-operator-1-build\" (UID: \"2f5ad768-60c4-430b-85c1-90a0f8cddc62\") " pod="service-telemetry/smart-gateway-operator-1-build" Dec 03 00:24:57 crc kubenswrapper[4953]: I1203 00:24:57.415149 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/2f5ad768-60c4-430b-85c1-90a0f8cddc62-build-ca-bundles\") pod \"smart-gateway-operator-1-build\" (UID: \"2f5ad768-60c4-430b-85c1-90a0f8cddc62\") " pod="service-telemetry/smart-gateway-operator-1-build" Dec 03 00:24:57 crc kubenswrapper[4953]: I1203 00:24:57.415334 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/2f5ad768-60c4-430b-85c1-90a0f8cddc62-container-storage-root\") pod \"smart-gateway-operator-1-build\" (UID: \"2f5ad768-60c4-430b-85c1-90a0f8cddc62\") " pod="service-telemetry/smart-gateway-operator-1-build" Dec 03 00:24:57 crc kubenswrapper[4953]: I1203 00:24:57.415522 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/2f5ad768-60c4-430b-85c1-90a0f8cddc62-build-blob-cache\") pod \"smart-gateway-operator-1-build\" (UID: \"2f5ad768-60c4-430b-85c1-90a0f8cddc62\") " pod="service-telemetry/smart-gateway-operator-1-build" Dec 03 00:24:57 crc kubenswrapper[4953]: I1203 00:24:57.415657 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/2f5ad768-60c4-430b-85c1-90a0f8cddc62-buildworkdir\") pod \"smart-gateway-operator-1-build\" (UID: \"2f5ad768-60c4-430b-85c1-90a0f8cddc62\") " pod="service-telemetry/smart-gateway-operator-1-build" Dec 03 00:24:57 crc kubenswrapper[4953]: I1203 00:24:57.416004 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"builder-dockercfg-9prwd-push\" (UniqueName: \"kubernetes.io/secret/2f5ad768-60c4-430b-85c1-90a0f8cddc62-builder-dockercfg-9prwd-push\") pod \"smart-gateway-operator-1-build\" (UID: \"2f5ad768-60c4-430b-85c1-90a0f8cddc62\") " pod="service-telemetry/smart-gateway-operator-1-build" Dec 03 00:24:57 crc kubenswrapper[4953]: I1203 00:24:57.416113 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/2f5ad768-60c4-430b-85c1-90a0f8cddc62-build-proxy-ca-bundles\") pod \"smart-gateway-operator-1-build\" (UID: \"2f5ad768-60c4-430b-85c1-90a0f8cddc62\") " pod="service-telemetry/smart-gateway-operator-1-build" Dec 03 00:24:57 crc kubenswrapper[4953]: I1203 00:24:57.416228 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/2f5ad768-60c4-430b-85c1-90a0f8cddc62-build-system-configs\") pod \"smart-gateway-operator-1-build\" (UID: \"2f5ad768-60c4-430b-85c1-90a0f8cddc62\") " pod="service-telemetry/smart-gateway-operator-1-build" Dec 03 00:24:57 crc kubenswrapper[4953]: I1203 00:24:57.416270 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/2f5ad768-60c4-430b-85c1-90a0f8cddc62-node-pullsecrets\") pod \"smart-gateway-operator-1-build\" (UID: \"2f5ad768-60c4-430b-85c1-90a0f8cddc62\") " pod="service-telemetry/smart-gateway-operator-1-build" Dec 03 00:24:57 crc kubenswrapper[4953]: I1203 00:24:57.416319 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/2f5ad768-60c4-430b-85c1-90a0f8cddc62-buildcachedir\") pod \"smart-gateway-operator-1-build\" (UID: \"2f5ad768-60c4-430b-85c1-90a0f8cddc62\") " pod="service-telemetry/smart-gateway-operator-1-build" Dec 03 00:24:57 crc kubenswrapper[4953]: I1203 00:24:57.416423 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/2f5ad768-60c4-430b-85c1-90a0f8cddc62-container-storage-run\") pod \"smart-gateway-operator-1-build\" (UID: \"2f5ad768-60c4-430b-85c1-90a0f8cddc62\") " pod="service-telemetry/smart-gateway-operator-1-build" Dec 03 00:24:57 crc kubenswrapper[4953]: I1203 00:24:57.416524 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v4m2l\" (UniqueName: \"kubernetes.io/projected/2f5ad768-60c4-430b-85c1-90a0f8cddc62-kube-api-access-v4m2l\") pod \"smart-gateway-operator-1-build\" (UID: \"2f5ad768-60c4-430b-85c1-90a0f8cddc62\") " pod="service-telemetry/smart-gateway-operator-1-build" Dec 03 00:24:57 crc kubenswrapper[4953]: I1203 00:24:57.518018 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"builder-dockercfg-9prwd-push\" (UniqueName: \"kubernetes.io/secret/2f5ad768-60c4-430b-85c1-90a0f8cddc62-builder-dockercfg-9prwd-push\") pod \"smart-gateway-operator-1-build\" (UID: \"2f5ad768-60c4-430b-85c1-90a0f8cddc62\") " pod="service-telemetry/smart-gateway-operator-1-build" Dec 03 00:24:57 crc kubenswrapper[4953]: I1203 00:24:57.518128 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/2f5ad768-60c4-430b-85c1-90a0f8cddc62-build-proxy-ca-bundles\") pod \"smart-gateway-operator-1-build\" (UID: \"2f5ad768-60c4-430b-85c1-90a0f8cddc62\") " pod="service-telemetry/smart-gateway-operator-1-build" Dec 03 00:24:57 crc kubenswrapper[4953]: I1203 00:24:57.518201 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/2f5ad768-60c4-430b-85c1-90a0f8cddc62-build-system-configs\") pod \"smart-gateway-operator-1-build\" (UID: \"2f5ad768-60c4-430b-85c1-90a0f8cddc62\") " pod="service-telemetry/smart-gateway-operator-1-build" Dec 03 00:24:57 crc kubenswrapper[4953]: I1203 00:24:57.518265 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/2f5ad768-60c4-430b-85c1-90a0f8cddc62-node-pullsecrets\") pod \"smart-gateway-operator-1-build\" (UID: \"2f5ad768-60c4-430b-85c1-90a0f8cddc62\") " pod="service-telemetry/smart-gateway-operator-1-build" Dec 03 00:24:57 crc kubenswrapper[4953]: I1203 00:24:57.518315 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/2f5ad768-60c4-430b-85c1-90a0f8cddc62-buildcachedir\") pod \"smart-gateway-operator-1-build\" (UID: \"2f5ad768-60c4-430b-85c1-90a0f8cddc62\") " pod="service-telemetry/smart-gateway-operator-1-build" Dec 03 00:24:57 crc kubenswrapper[4953]: I1203 00:24:57.518375 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/2f5ad768-60c4-430b-85c1-90a0f8cddc62-container-storage-run\") pod \"smart-gateway-operator-1-build\" (UID: \"2f5ad768-60c4-430b-85c1-90a0f8cddc62\") " pod="service-telemetry/smart-gateway-operator-1-build" Dec 03 00:24:57 crc kubenswrapper[4953]: I1203 00:24:57.518430 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v4m2l\" (UniqueName: \"kubernetes.io/projected/2f5ad768-60c4-430b-85c1-90a0f8cddc62-kube-api-access-v4m2l\") pod \"smart-gateway-operator-1-build\" (UID: \"2f5ad768-60c4-430b-85c1-90a0f8cddc62\") " pod="service-telemetry/smart-gateway-operator-1-build" Dec 03 00:24:57 crc kubenswrapper[4953]: I1203 00:24:57.518526 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"builder-dockercfg-9prwd-pull\" (UniqueName: \"kubernetes.io/secret/2f5ad768-60c4-430b-85c1-90a0f8cddc62-builder-dockercfg-9prwd-pull\") pod \"smart-gateway-operator-1-build\" (UID: \"2f5ad768-60c4-430b-85c1-90a0f8cddc62\") " pod="service-telemetry/smart-gateway-operator-1-build" Dec 03 00:24:57 crc kubenswrapper[4953]: I1203 00:24:57.518547 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/2f5ad768-60c4-430b-85c1-90a0f8cddc62-node-pullsecrets\") pod \"smart-gateway-operator-1-build\" (UID: \"2f5ad768-60c4-430b-85c1-90a0f8cddc62\") " pod="service-telemetry/smart-gateway-operator-1-build" Dec 03 00:24:57 crc kubenswrapper[4953]: I1203 00:24:57.518666 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/2f5ad768-60c4-430b-85c1-90a0f8cddc62-buildcachedir\") pod \"smart-gateway-operator-1-build\" (UID: \"2f5ad768-60c4-430b-85c1-90a0f8cddc62\") " pod="service-telemetry/smart-gateway-operator-1-build" Dec 03 00:24:57 crc kubenswrapper[4953]: I1203 00:24:57.519459 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/2f5ad768-60c4-430b-85c1-90a0f8cddc62-container-storage-run\") pod \"smart-gateway-operator-1-build\" (UID: \"2f5ad768-60c4-430b-85c1-90a0f8cddc62\") " pod="service-telemetry/smart-gateway-operator-1-build" Dec 03 00:24:57 crc kubenswrapper[4953]: I1203 00:24:57.518582 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/2f5ad768-60c4-430b-85c1-90a0f8cddc62-build-ca-bundles\") pod \"smart-gateway-operator-1-build\" (UID: \"2f5ad768-60c4-430b-85c1-90a0f8cddc62\") " pod="service-telemetry/smart-gateway-operator-1-build" Dec 03 00:24:57 crc kubenswrapper[4953]: I1203 00:24:57.519818 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/2f5ad768-60c4-430b-85c1-90a0f8cddc62-build-system-configs\") pod \"smart-gateway-operator-1-build\" (UID: \"2f5ad768-60c4-430b-85c1-90a0f8cddc62\") " pod="service-telemetry/smart-gateway-operator-1-build" Dec 03 00:24:57 crc kubenswrapper[4953]: I1203 00:24:57.519822 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/2f5ad768-60c4-430b-85c1-90a0f8cddc62-container-storage-root\") pod \"smart-gateway-operator-1-build\" (UID: \"2f5ad768-60c4-430b-85c1-90a0f8cddc62\") " pod="service-telemetry/smart-gateway-operator-1-build" Dec 03 00:24:57 crc kubenswrapper[4953]: I1203 00:24:57.520035 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/2f5ad768-60c4-430b-85c1-90a0f8cddc62-build-ca-bundles\") pod \"smart-gateway-operator-1-build\" (UID: \"2f5ad768-60c4-430b-85c1-90a0f8cddc62\") " pod="service-telemetry/smart-gateway-operator-1-build" Dec 03 00:24:57 crc kubenswrapper[4953]: I1203 00:24:57.519662 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/2f5ad768-60c4-430b-85c1-90a0f8cddc62-build-proxy-ca-bundles\") pod \"smart-gateway-operator-1-build\" (UID: \"2f5ad768-60c4-430b-85c1-90a0f8cddc62\") " pod="service-telemetry/smart-gateway-operator-1-build" Dec 03 00:24:57 crc kubenswrapper[4953]: I1203 00:24:57.520282 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/2f5ad768-60c4-430b-85c1-90a0f8cddc62-container-storage-root\") pod \"smart-gateway-operator-1-build\" (UID: \"2f5ad768-60c4-430b-85c1-90a0f8cddc62\") " pod="service-telemetry/smart-gateway-operator-1-build" Dec 03 00:24:57 crc kubenswrapper[4953]: I1203 00:24:57.520434 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/2f5ad768-60c4-430b-85c1-90a0f8cddc62-build-blob-cache\") pod \"smart-gateway-operator-1-build\" (UID: \"2f5ad768-60c4-430b-85c1-90a0f8cddc62\") " pod="service-telemetry/smart-gateway-operator-1-build" Dec 03 00:24:57 crc kubenswrapper[4953]: I1203 00:24:57.520491 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/2f5ad768-60c4-430b-85c1-90a0f8cddc62-buildworkdir\") pod \"smart-gateway-operator-1-build\" (UID: \"2f5ad768-60c4-430b-85c1-90a0f8cddc62\") " pod="service-telemetry/smart-gateway-operator-1-build" Dec 03 00:24:57 crc kubenswrapper[4953]: I1203 00:24:57.520894 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/2f5ad768-60c4-430b-85c1-90a0f8cddc62-buildworkdir\") pod \"smart-gateway-operator-1-build\" (UID: \"2f5ad768-60c4-430b-85c1-90a0f8cddc62\") " pod="service-telemetry/smart-gateway-operator-1-build" Dec 03 00:24:57 crc kubenswrapper[4953]: I1203 00:24:57.521404 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/2f5ad768-60c4-430b-85c1-90a0f8cddc62-build-blob-cache\") pod \"smart-gateway-operator-1-build\" (UID: \"2f5ad768-60c4-430b-85c1-90a0f8cddc62\") " pod="service-telemetry/smart-gateway-operator-1-build" Dec 03 00:24:57 crc kubenswrapper[4953]: I1203 00:24:57.526613 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"builder-dockercfg-9prwd-pull\" (UniqueName: \"kubernetes.io/secret/2f5ad768-60c4-430b-85c1-90a0f8cddc62-builder-dockercfg-9prwd-pull\") pod \"smart-gateway-operator-1-build\" (UID: \"2f5ad768-60c4-430b-85c1-90a0f8cddc62\") " pod="service-telemetry/smart-gateway-operator-1-build" Dec 03 00:24:57 crc kubenswrapper[4953]: I1203 00:24:57.528264 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"builder-dockercfg-9prwd-push\" (UniqueName: \"kubernetes.io/secret/2f5ad768-60c4-430b-85c1-90a0f8cddc62-builder-dockercfg-9prwd-push\") pod \"smart-gateway-operator-1-build\" (UID: \"2f5ad768-60c4-430b-85c1-90a0f8cddc62\") " pod="service-telemetry/smart-gateway-operator-1-build" Dec 03 00:24:57 crc kubenswrapper[4953]: I1203 00:24:57.541162 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v4m2l\" (UniqueName: \"kubernetes.io/projected/2f5ad768-60c4-430b-85c1-90a0f8cddc62-kube-api-access-v4m2l\") pod \"smart-gateway-operator-1-build\" (UID: \"2f5ad768-60c4-430b-85c1-90a0f8cddc62\") " pod="service-telemetry/smart-gateway-operator-1-build" Dec 03 00:24:57 crc kubenswrapper[4953]: I1203 00:24:57.664589 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/smart-gateway-operator-1-build" Dec 03 00:24:57 crc kubenswrapper[4953]: I1203 00:24:57.891562 4953 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/smart-gateway-operator-1-build"] Dec 03 00:24:58 crc kubenswrapper[4953]: I1203 00:24:58.100880 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/smart-gateway-operator-1-build" event={"ID":"2f5ad768-60c4-430b-85c1-90a0f8cddc62","Type":"ContainerStarted","Data":"5c4c8d26b24eaec1c9dbd304158cd9e86dd754b45d5e5051f856f03f550c7873"} Dec 03 00:24:59 crc kubenswrapper[4953]: I1203 00:24:59.114068 4953 generic.go:334] "Generic (PLEG): container finished" podID="2f5ad768-60c4-430b-85c1-90a0f8cddc62" containerID="24ef1cb9f22e5dba721a2f96e5ec39c2559422a67ff70d99e0fff3724af55b5a" exitCode=0 Dec 03 00:24:59 crc kubenswrapper[4953]: I1203 00:24:59.114171 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/smart-gateway-operator-1-build" event={"ID":"2f5ad768-60c4-430b-85c1-90a0f8cddc62","Type":"ContainerDied","Data":"24ef1cb9f22e5dba721a2f96e5ec39c2559422a67ff70d99e0fff3724af55b5a"} Dec 03 00:25:00 crc kubenswrapper[4953]: I1203 00:25:00.123830 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/smart-gateway-operator-1-build" event={"ID":"2f5ad768-60c4-430b-85c1-90a0f8cddc62","Type":"ContainerStarted","Data":"f81776546e6a27a3ab999ecd53085aa3aaf11a750cdfc71dce39d55017cb87b3"} Dec 03 00:25:00 crc kubenswrapper[4953]: I1203 00:25:00.155377 4953 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="service-telemetry/smart-gateway-operator-1-build" podStartSLOduration=3.15534856 podStartE2EDuration="3.15534856s" podCreationTimestamp="2025-12-03 00:24:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 00:25:00.153932545 +0000 UTC m=+1089.138213498" watchObservedRunningTime="2025-12-03 00:25:00.15534856 +0000 UTC m=+1089.139629343" Dec 03 00:25:08 crc kubenswrapper[4953]: I1203 00:25:08.125872 4953 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["service-telemetry/smart-gateway-operator-1-build"] Dec 03 00:25:08 crc kubenswrapper[4953]: I1203 00:25:08.127452 4953 kuberuntime_container.go:808] "Killing container with a grace period" pod="service-telemetry/smart-gateway-operator-1-build" podUID="2f5ad768-60c4-430b-85c1-90a0f8cddc62" containerName="docker-build" containerID="cri-o://f81776546e6a27a3ab999ecd53085aa3aaf11a750cdfc71dce39d55017cb87b3" gracePeriod=30 Dec 03 00:25:09 crc kubenswrapper[4953]: I1203 00:25:09.203553 4953 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_smart-gateway-operator-1-build_2f5ad768-60c4-430b-85c1-90a0f8cddc62/docker-build/0.log" Dec 03 00:25:09 crc kubenswrapper[4953]: I1203 00:25:09.204510 4953 generic.go:334] "Generic (PLEG): container finished" podID="2f5ad768-60c4-430b-85c1-90a0f8cddc62" containerID="f81776546e6a27a3ab999ecd53085aa3aaf11a750cdfc71dce39d55017cb87b3" exitCode=1 Dec 03 00:25:09 crc kubenswrapper[4953]: I1203 00:25:09.204563 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/smart-gateway-operator-1-build" event={"ID":"2f5ad768-60c4-430b-85c1-90a0f8cddc62","Type":"ContainerDied","Data":"f81776546e6a27a3ab999ecd53085aa3aaf11a750cdfc71dce39d55017cb87b3"} Dec 03 00:25:09 crc kubenswrapper[4953]: I1203 00:25:09.328147 4953 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_smart-gateway-operator-1-build_2f5ad768-60c4-430b-85c1-90a0f8cddc62/docker-build/0.log" Dec 03 00:25:09 crc kubenswrapper[4953]: I1203 00:25:09.328614 4953 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/smart-gateway-operator-1-build" Dec 03 00:25:09 crc kubenswrapper[4953]: I1203 00:25:09.464121 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v4m2l\" (UniqueName: \"kubernetes.io/projected/2f5ad768-60c4-430b-85c1-90a0f8cddc62-kube-api-access-v4m2l\") pod \"2f5ad768-60c4-430b-85c1-90a0f8cddc62\" (UID: \"2f5ad768-60c4-430b-85c1-90a0f8cddc62\") " Dec 03 00:25:09 crc kubenswrapper[4953]: I1203 00:25:09.464236 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/2f5ad768-60c4-430b-85c1-90a0f8cddc62-build-blob-cache\") pod \"2f5ad768-60c4-430b-85c1-90a0f8cddc62\" (UID: \"2f5ad768-60c4-430b-85c1-90a0f8cddc62\") " Dec 03 00:25:09 crc kubenswrapper[4953]: I1203 00:25:09.464273 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/2f5ad768-60c4-430b-85c1-90a0f8cddc62-buildcachedir\") pod \"2f5ad768-60c4-430b-85c1-90a0f8cddc62\" (UID: \"2f5ad768-60c4-430b-85c1-90a0f8cddc62\") " Dec 03 00:25:09 crc kubenswrapper[4953]: I1203 00:25:09.464295 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/2f5ad768-60c4-430b-85c1-90a0f8cddc62-node-pullsecrets\") pod \"2f5ad768-60c4-430b-85c1-90a0f8cddc62\" (UID: \"2f5ad768-60c4-430b-85c1-90a0f8cddc62\") " Dec 03 00:25:09 crc kubenswrapper[4953]: I1203 00:25:09.464362 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/2f5ad768-60c4-430b-85c1-90a0f8cddc62-container-storage-root\") pod \"2f5ad768-60c4-430b-85c1-90a0f8cddc62\" (UID: \"2f5ad768-60c4-430b-85c1-90a0f8cddc62\") " Dec 03 00:25:09 crc kubenswrapper[4953]: I1203 00:25:09.464410 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/2f5ad768-60c4-430b-85c1-90a0f8cddc62-build-system-configs\") pod \"2f5ad768-60c4-430b-85c1-90a0f8cddc62\" (UID: \"2f5ad768-60c4-430b-85c1-90a0f8cddc62\") " Dec 03 00:25:09 crc kubenswrapper[4953]: I1203 00:25:09.464439 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/2f5ad768-60c4-430b-85c1-90a0f8cddc62-build-proxy-ca-bundles\") pod \"2f5ad768-60c4-430b-85c1-90a0f8cddc62\" (UID: \"2f5ad768-60c4-430b-85c1-90a0f8cddc62\") " Dec 03 00:25:09 crc kubenswrapper[4953]: I1203 00:25:09.464506 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"builder-dockercfg-9prwd-pull\" (UniqueName: \"kubernetes.io/secret/2f5ad768-60c4-430b-85c1-90a0f8cddc62-builder-dockercfg-9prwd-pull\") pod \"2f5ad768-60c4-430b-85c1-90a0f8cddc62\" (UID: \"2f5ad768-60c4-430b-85c1-90a0f8cddc62\") " Dec 03 00:25:09 crc kubenswrapper[4953]: I1203 00:25:09.464487 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2f5ad768-60c4-430b-85c1-90a0f8cddc62-node-pullsecrets" (OuterVolumeSpecName: "node-pullsecrets") pod "2f5ad768-60c4-430b-85c1-90a0f8cddc62" (UID: "2f5ad768-60c4-430b-85c1-90a0f8cddc62"). InnerVolumeSpecName "node-pullsecrets". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 00:25:09 crc kubenswrapper[4953]: I1203 00:25:09.464535 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"builder-dockercfg-9prwd-push\" (UniqueName: \"kubernetes.io/secret/2f5ad768-60c4-430b-85c1-90a0f8cddc62-builder-dockercfg-9prwd-push\") pod \"2f5ad768-60c4-430b-85c1-90a0f8cddc62\" (UID: \"2f5ad768-60c4-430b-85c1-90a0f8cddc62\") " Dec 03 00:25:09 crc kubenswrapper[4953]: I1203 00:25:09.464611 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/2f5ad768-60c4-430b-85c1-90a0f8cddc62-container-storage-run\") pod \"2f5ad768-60c4-430b-85c1-90a0f8cddc62\" (UID: \"2f5ad768-60c4-430b-85c1-90a0f8cddc62\") " Dec 03 00:25:09 crc kubenswrapper[4953]: I1203 00:25:09.464631 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2f5ad768-60c4-430b-85c1-90a0f8cddc62-buildcachedir" (OuterVolumeSpecName: "buildcachedir") pod "2f5ad768-60c4-430b-85c1-90a0f8cddc62" (UID: "2f5ad768-60c4-430b-85c1-90a0f8cddc62"). InnerVolumeSpecName "buildcachedir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 00:25:09 crc kubenswrapper[4953]: I1203 00:25:09.464654 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/2f5ad768-60c4-430b-85c1-90a0f8cddc62-build-ca-bundles\") pod \"2f5ad768-60c4-430b-85c1-90a0f8cddc62\" (UID: \"2f5ad768-60c4-430b-85c1-90a0f8cddc62\") " Dec 03 00:25:09 crc kubenswrapper[4953]: I1203 00:25:09.464673 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/2f5ad768-60c4-430b-85c1-90a0f8cddc62-buildworkdir\") pod \"2f5ad768-60c4-430b-85c1-90a0f8cddc62\" (UID: \"2f5ad768-60c4-430b-85c1-90a0f8cddc62\") " Dec 03 00:25:09 crc kubenswrapper[4953]: I1203 00:25:09.465028 4953 reconciler_common.go:293] "Volume detached for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/2f5ad768-60c4-430b-85c1-90a0f8cddc62-buildcachedir\") on node \"crc\" DevicePath \"\"" Dec 03 00:25:09 crc kubenswrapper[4953]: I1203 00:25:09.465044 4953 reconciler_common.go:293] "Volume detached for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/2f5ad768-60c4-430b-85c1-90a0f8cddc62-node-pullsecrets\") on node \"crc\" DevicePath \"\"" Dec 03 00:25:09 crc kubenswrapper[4953]: I1203 00:25:09.465428 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2f5ad768-60c4-430b-85c1-90a0f8cddc62-build-proxy-ca-bundles" (OuterVolumeSpecName: "build-proxy-ca-bundles") pod "2f5ad768-60c4-430b-85c1-90a0f8cddc62" (UID: "2f5ad768-60c4-430b-85c1-90a0f8cddc62"). InnerVolumeSpecName "build-proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:25:09 crc kubenswrapper[4953]: I1203 00:25:09.465837 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2f5ad768-60c4-430b-85c1-90a0f8cddc62-build-system-configs" (OuterVolumeSpecName: "build-system-configs") pod "2f5ad768-60c4-430b-85c1-90a0f8cddc62" (UID: "2f5ad768-60c4-430b-85c1-90a0f8cddc62"). InnerVolumeSpecName "build-system-configs". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:25:09 crc kubenswrapper[4953]: I1203 00:25:09.466914 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2f5ad768-60c4-430b-85c1-90a0f8cddc62-buildworkdir" (OuterVolumeSpecName: "buildworkdir") pod "2f5ad768-60c4-430b-85c1-90a0f8cddc62" (UID: "2f5ad768-60c4-430b-85c1-90a0f8cddc62"). InnerVolumeSpecName "buildworkdir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 00:25:09 crc kubenswrapper[4953]: I1203 00:25:09.467210 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2f5ad768-60c4-430b-85c1-90a0f8cddc62-build-ca-bundles" (OuterVolumeSpecName: "build-ca-bundles") pod "2f5ad768-60c4-430b-85c1-90a0f8cddc62" (UID: "2f5ad768-60c4-430b-85c1-90a0f8cddc62"). InnerVolumeSpecName "build-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:25:09 crc kubenswrapper[4953]: I1203 00:25:09.467404 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2f5ad768-60c4-430b-85c1-90a0f8cddc62-container-storage-run" (OuterVolumeSpecName: "container-storage-run") pod "2f5ad768-60c4-430b-85c1-90a0f8cddc62" (UID: "2f5ad768-60c4-430b-85c1-90a0f8cddc62"). InnerVolumeSpecName "container-storage-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 00:25:09 crc kubenswrapper[4953]: I1203 00:25:09.468199 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2f5ad768-60c4-430b-85c1-90a0f8cddc62-container-storage-root" (OuterVolumeSpecName: "container-storage-root") pod "2f5ad768-60c4-430b-85c1-90a0f8cddc62" (UID: "2f5ad768-60c4-430b-85c1-90a0f8cddc62"). InnerVolumeSpecName "container-storage-root". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 00:25:09 crc kubenswrapper[4953]: I1203 00:25:09.473709 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2f5ad768-60c4-430b-85c1-90a0f8cddc62-builder-dockercfg-9prwd-push" (OuterVolumeSpecName: "builder-dockercfg-9prwd-push") pod "2f5ad768-60c4-430b-85c1-90a0f8cddc62" (UID: "2f5ad768-60c4-430b-85c1-90a0f8cddc62"). InnerVolumeSpecName "builder-dockercfg-9prwd-push". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:25:09 crc kubenswrapper[4953]: I1203 00:25:09.473897 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2f5ad768-60c4-430b-85c1-90a0f8cddc62-builder-dockercfg-9prwd-pull" (OuterVolumeSpecName: "builder-dockercfg-9prwd-pull") pod "2f5ad768-60c4-430b-85c1-90a0f8cddc62" (UID: "2f5ad768-60c4-430b-85c1-90a0f8cddc62"). InnerVolumeSpecName "builder-dockercfg-9prwd-pull". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:25:09 crc kubenswrapper[4953]: I1203 00:25:09.475057 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2f5ad768-60c4-430b-85c1-90a0f8cddc62-kube-api-access-v4m2l" (OuterVolumeSpecName: "kube-api-access-v4m2l") pod "2f5ad768-60c4-430b-85c1-90a0f8cddc62" (UID: "2f5ad768-60c4-430b-85c1-90a0f8cddc62"). InnerVolumeSpecName "kube-api-access-v4m2l". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:25:09 crc kubenswrapper[4953]: I1203 00:25:09.566595 4953 reconciler_common.go:293] "Volume detached for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/2f5ad768-60c4-430b-85c1-90a0f8cddc62-container-storage-root\") on node \"crc\" DevicePath \"\"" Dec 03 00:25:09 crc kubenswrapper[4953]: I1203 00:25:09.566644 4953 reconciler_common.go:293] "Volume detached for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/2f5ad768-60c4-430b-85c1-90a0f8cddc62-build-system-configs\") on node \"crc\" DevicePath \"\"" Dec 03 00:25:09 crc kubenswrapper[4953]: I1203 00:25:09.566658 4953 reconciler_common.go:293] "Volume detached for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/2f5ad768-60c4-430b-85c1-90a0f8cddc62-build-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Dec 03 00:25:09 crc kubenswrapper[4953]: I1203 00:25:09.566671 4953 reconciler_common.go:293] "Volume detached for volume \"builder-dockercfg-9prwd-pull\" (UniqueName: \"kubernetes.io/secret/2f5ad768-60c4-430b-85c1-90a0f8cddc62-builder-dockercfg-9prwd-pull\") on node \"crc\" DevicePath \"\"" Dec 03 00:25:09 crc kubenswrapper[4953]: I1203 00:25:09.566683 4953 reconciler_common.go:293] "Volume detached for volume \"builder-dockercfg-9prwd-push\" (UniqueName: \"kubernetes.io/secret/2f5ad768-60c4-430b-85c1-90a0f8cddc62-builder-dockercfg-9prwd-push\") on node \"crc\" DevicePath \"\"" Dec 03 00:25:09 crc kubenswrapper[4953]: I1203 00:25:09.566695 4953 reconciler_common.go:293] "Volume detached for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/2f5ad768-60c4-430b-85c1-90a0f8cddc62-container-storage-run\") on node \"crc\" DevicePath \"\"" Dec 03 00:25:09 crc kubenswrapper[4953]: I1203 00:25:09.566706 4953 reconciler_common.go:293] "Volume detached for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/2f5ad768-60c4-430b-85c1-90a0f8cddc62-build-ca-bundles\") on node \"crc\" DevicePath \"\"" Dec 03 00:25:09 crc kubenswrapper[4953]: I1203 00:25:09.566723 4953 reconciler_common.go:293] "Volume detached for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/2f5ad768-60c4-430b-85c1-90a0f8cddc62-buildworkdir\") on node \"crc\" DevicePath \"\"" Dec 03 00:25:09 crc kubenswrapper[4953]: I1203 00:25:09.566736 4953 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v4m2l\" (UniqueName: \"kubernetes.io/projected/2f5ad768-60c4-430b-85c1-90a0f8cddc62-kube-api-access-v4m2l\") on node \"crc\" DevicePath \"\"" Dec 03 00:25:09 crc kubenswrapper[4953]: I1203 00:25:09.685072 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2f5ad768-60c4-430b-85c1-90a0f8cddc62-build-blob-cache" (OuterVolumeSpecName: "build-blob-cache") pod "2f5ad768-60c4-430b-85c1-90a0f8cddc62" (UID: "2f5ad768-60c4-430b-85c1-90a0f8cddc62"). InnerVolumeSpecName "build-blob-cache". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 00:25:09 crc kubenswrapper[4953]: I1203 00:25:09.716790 4953 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["service-telemetry/smart-gateway-operator-2-build"] Dec 03 00:25:09 crc kubenswrapper[4953]: E1203 00:25:09.717229 4953 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2f5ad768-60c4-430b-85c1-90a0f8cddc62" containerName="docker-build" Dec 03 00:25:09 crc kubenswrapper[4953]: I1203 00:25:09.717253 4953 state_mem.go:107] "Deleted CPUSet assignment" podUID="2f5ad768-60c4-430b-85c1-90a0f8cddc62" containerName="docker-build" Dec 03 00:25:09 crc kubenswrapper[4953]: E1203 00:25:09.717269 4953 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2f5ad768-60c4-430b-85c1-90a0f8cddc62" containerName="manage-dockerfile" Dec 03 00:25:09 crc kubenswrapper[4953]: I1203 00:25:09.717279 4953 state_mem.go:107] "Deleted CPUSet assignment" podUID="2f5ad768-60c4-430b-85c1-90a0f8cddc62" containerName="manage-dockerfile" Dec 03 00:25:09 crc kubenswrapper[4953]: I1203 00:25:09.717435 4953 memory_manager.go:354] "RemoveStaleState removing state" podUID="2f5ad768-60c4-430b-85c1-90a0f8cddc62" containerName="docker-build" Dec 03 00:25:09 crc kubenswrapper[4953]: I1203 00:25:09.719386 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/smart-gateway-operator-2-build" Dec 03 00:25:09 crc kubenswrapper[4953]: I1203 00:25:09.721843 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"smart-gateway-operator-2-global-ca" Dec 03 00:25:09 crc kubenswrapper[4953]: I1203 00:25:09.722159 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"smart-gateway-operator-2-sys-config" Dec 03 00:25:09 crc kubenswrapper[4953]: I1203 00:25:09.722410 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"smart-gateway-operator-2-ca" Dec 03 00:25:09 crc kubenswrapper[4953]: I1203 00:25:09.742233 4953 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/smart-gateway-operator-2-build"] Dec 03 00:25:09 crc kubenswrapper[4953]: I1203 00:25:09.787817 4953 reconciler_common.go:293] "Volume detached for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/2f5ad768-60c4-430b-85c1-90a0f8cddc62-build-blob-cache\") on node \"crc\" DevicePath \"\"" Dec 03 00:25:09 crc kubenswrapper[4953]: I1203 00:25:09.889245 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/492fe3e3-8abb-40ef-a0ea-93334e6fe94e-buildworkdir\") pod \"smart-gateway-operator-2-build\" (UID: \"492fe3e3-8abb-40ef-a0ea-93334e6fe94e\") " pod="service-telemetry/smart-gateway-operator-2-build" Dec 03 00:25:09 crc kubenswrapper[4953]: I1203 00:25:09.889308 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/492fe3e3-8abb-40ef-a0ea-93334e6fe94e-node-pullsecrets\") pod \"smart-gateway-operator-2-build\" (UID: \"492fe3e3-8abb-40ef-a0ea-93334e6fe94e\") " pod="service-telemetry/smart-gateway-operator-2-build" Dec 03 00:25:09 crc kubenswrapper[4953]: I1203 00:25:09.889350 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/492fe3e3-8abb-40ef-a0ea-93334e6fe94e-build-blob-cache\") pod \"smart-gateway-operator-2-build\" (UID: \"492fe3e3-8abb-40ef-a0ea-93334e6fe94e\") " pod="service-telemetry/smart-gateway-operator-2-build" Dec 03 00:25:09 crc kubenswrapper[4953]: I1203 00:25:09.889411 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/492fe3e3-8abb-40ef-a0ea-93334e6fe94e-build-ca-bundles\") pod \"smart-gateway-operator-2-build\" (UID: \"492fe3e3-8abb-40ef-a0ea-93334e6fe94e\") " pod="service-telemetry/smart-gateway-operator-2-build" Dec 03 00:25:09 crc kubenswrapper[4953]: I1203 00:25:09.889446 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/492fe3e3-8abb-40ef-a0ea-93334e6fe94e-container-storage-run\") pod \"smart-gateway-operator-2-build\" (UID: \"492fe3e3-8abb-40ef-a0ea-93334e6fe94e\") " pod="service-telemetry/smart-gateway-operator-2-build" Dec 03 00:25:09 crc kubenswrapper[4953]: I1203 00:25:09.889474 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"builder-dockercfg-9prwd-push\" (UniqueName: \"kubernetes.io/secret/492fe3e3-8abb-40ef-a0ea-93334e6fe94e-builder-dockercfg-9prwd-push\") pod \"smart-gateway-operator-2-build\" (UID: \"492fe3e3-8abb-40ef-a0ea-93334e6fe94e\") " pod="service-telemetry/smart-gateway-operator-2-build" Dec 03 00:25:09 crc kubenswrapper[4953]: I1203 00:25:09.889511 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/492fe3e3-8abb-40ef-a0ea-93334e6fe94e-build-proxy-ca-bundles\") pod \"smart-gateway-operator-2-build\" (UID: \"492fe3e3-8abb-40ef-a0ea-93334e6fe94e\") " pod="service-telemetry/smart-gateway-operator-2-build" Dec 03 00:25:09 crc kubenswrapper[4953]: I1203 00:25:09.889541 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/492fe3e3-8abb-40ef-a0ea-93334e6fe94e-buildcachedir\") pod \"smart-gateway-operator-2-build\" (UID: \"492fe3e3-8abb-40ef-a0ea-93334e6fe94e\") " pod="service-telemetry/smart-gateway-operator-2-build" Dec 03 00:25:09 crc kubenswrapper[4953]: I1203 00:25:09.889581 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/492fe3e3-8abb-40ef-a0ea-93334e6fe94e-build-system-configs\") pod \"smart-gateway-operator-2-build\" (UID: \"492fe3e3-8abb-40ef-a0ea-93334e6fe94e\") " pod="service-telemetry/smart-gateway-operator-2-build" Dec 03 00:25:09 crc kubenswrapper[4953]: I1203 00:25:09.889608 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sv78b\" (UniqueName: \"kubernetes.io/projected/492fe3e3-8abb-40ef-a0ea-93334e6fe94e-kube-api-access-sv78b\") pod \"smart-gateway-operator-2-build\" (UID: \"492fe3e3-8abb-40ef-a0ea-93334e6fe94e\") " pod="service-telemetry/smart-gateway-operator-2-build" Dec 03 00:25:09 crc kubenswrapper[4953]: I1203 00:25:09.889651 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"builder-dockercfg-9prwd-pull\" (UniqueName: \"kubernetes.io/secret/492fe3e3-8abb-40ef-a0ea-93334e6fe94e-builder-dockercfg-9prwd-pull\") pod \"smart-gateway-operator-2-build\" (UID: \"492fe3e3-8abb-40ef-a0ea-93334e6fe94e\") " pod="service-telemetry/smart-gateway-operator-2-build" Dec 03 00:25:09 crc kubenswrapper[4953]: I1203 00:25:09.889687 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/492fe3e3-8abb-40ef-a0ea-93334e6fe94e-container-storage-root\") pod \"smart-gateway-operator-2-build\" (UID: \"492fe3e3-8abb-40ef-a0ea-93334e6fe94e\") " pod="service-telemetry/smart-gateway-operator-2-build" Dec 03 00:25:09 crc kubenswrapper[4953]: I1203 00:25:09.991687 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/492fe3e3-8abb-40ef-a0ea-93334e6fe94e-build-proxy-ca-bundles\") pod \"smart-gateway-operator-2-build\" (UID: \"492fe3e3-8abb-40ef-a0ea-93334e6fe94e\") " pod="service-telemetry/smart-gateway-operator-2-build" Dec 03 00:25:09 crc kubenswrapper[4953]: I1203 00:25:09.991760 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/492fe3e3-8abb-40ef-a0ea-93334e6fe94e-buildcachedir\") pod \"smart-gateway-operator-2-build\" (UID: \"492fe3e3-8abb-40ef-a0ea-93334e6fe94e\") " pod="service-telemetry/smart-gateway-operator-2-build" Dec 03 00:25:09 crc kubenswrapper[4953]: I1203 00:25:09.991811 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/492fe3e3-8abb-40ef-a0ea-93334e6fe94e-build-system-configs\") pod \"smart-gateway-operator-2-build\" (UID: \"492fe3e3-8abb-40ef-a0ea-93334e6fe94e\") " pod="service-telemetry/smart-gateway-operator-2-build" Dec 03 00:25:09 crc kubenswrapper[4953]: I1203 00:25:09.991832 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"builder-dockercfg-9prwd-pull\" (UniqueName: \"kubernetes.io/secret/492fe3e3-8abb-40ef-a0ea-93334e6fe94e-builder-dockercfg-9prwd-pull\") pod \"smart-gateway-operator-2-build\" (UID: \"492fe3e3-8abb-40ef-a0ea-93334e6fe94e\") " pod="service-telemetry/smart-gateway-operator-2-build" Dec 03 00:25:09 crc kubenswrapper[4953]: I1203 00:25:09.991853 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sv78b\" (UniqueName: \"kubernetes.io/projected/492fe3e3-8abb-40ef-a0ea-93334e6fe94e-kube-api-access-sv78b\") pod \"smart-gateway-operator-2-build\" (UID: \"492fe3e3-8abb-40ef-a0ea-93334e6fe94e\") " pod="service-telemetry/smart-gateway-operator-2-build" Dec 03 00:25:09 crc kubenswrapper[4953]: I1203 00:25:09.991886 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/492fe3e3-8abb-40ef-a0ea-93334e6fe94e-container-storage-root\") pod \"smart-gateway-operator-2-build\" (UID: \"492fe3e3-8abb-40ef-a0ea-93334e6fe94e\") " pod="service-telemetry/smart-gateway-operator-2-build" Dec 03 00:25:09 crc kubenswrapper[4953]: I1203 00:25:09.991904 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/492fe3e3-8abb-40ef-a0ea-93334e6fe94e-node-pullsecrets\") pod \"smart-gateway-operator-2-build\" (UID: \"492fe3e3-8abb-40ef-a0ea-93334e6fe94e\") " pod="service-telemetry/smart-gateway-operator-2-build" Dec 03 00:25:09 crc kubenswrapper[4953]: I1203 00:25:09.991921 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/492fe3e3-8abb-40ef-a0ea-93334e6fe94e-buildworkdir\") pod \"smart-gateway-operator-2-build\" (UID: \"492fe3e3-8abb-40ef-a0ea-93334e6fe94e\") " pod="service-telemetry/smart-gateway-operator-2-build" Dec 03 00:25:09 crc kubenswrapper[4953]: I1203 00:25:09.991944 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/492fe3e3-8abb-40ef-a0ea-93334e6fe94e-build-blob-cache\") pod \"smart-gateway-operator-2-build\" (UID: \"492fe3e3-8abb-40ef-a0ea-93334e6fe94e\") " pod="service-telemetry/smart-gateway-operator-2-build" Dec 03 00:25:09 crc kubenswrapper[4953]: I1203 00:25:09.991999 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/492fe3e3-8abb-40ef-a0ea-93334e6fe94e-build-ca-bundles\") pod \"smart-gateway-operator-2-build\" (UID: \"492fe3e3-8abb-40ef-a0ea-93334e6fe94e\") " pod="service-telemetry/smart-gateway-operator-2-build" Dec 03 00:25:09 crc kubenswrapper[4953]: I1203 00:25:09.992041 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/492fe3e3-8abb-40ef-a0ea-93334e6fe94e-container-storage-run\") pod \"smart-gateway-operator-2-build\" (UID: \"492fe3e3-8abb-40ef-a0ea-93334e6fe94e\") " pod="service-telemetry/smart-gateway-operator-2-build" Dec 03 00:25:09 crc kubenswrapper[4953]: I1203 00:25:09.992069 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"builder-dockercfg-9prwd-push\" (UniqueName: \"kubernetes.io/secret/492fe3e3-8abb-40ef-a0ea-93334e6fe94e-builder-dockercfg-9prwd-push\") pod \"smart-gateway-operator-2-build\" (UID: \"492fe3e3-8abb-40ef-a0ea-93334e6fe94e\") " pod="service-telemetry/smart-gateway-operator-2-build" Dec 03 00:25:09 crc kubenswrapper[4953]: I1203 00:25:09.993434 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/492fe3e3-8abb-40ef-a0ea-93334e6fe94e-container-storage-root\") pod \"smart-gateway-operator-2-build\" (UID: \"492fe3e3-8abb-40ef-a0ea-93334e6fe94e\") " pod="service-telemetry/smart-gateway-operator-2-build" Dec 03 00:25:09 crc kubenswrapper[4953]: I1203 00:25:09.994121 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/492fe3e3-8abb-40ef-a0ea-93334e6fe94e-build-system-configs\") pod \"smart-gateway-operator-2-build\" (UID: \"492fe3e3-8abb-40ef-a0ea-93334e6fe94e\") " pod="service-telemetry/smart-gateway-operator-2-build" Dec 03 00:25:09 crc kubenswrapper[4953]: I1203 00:25:09.994753 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/492fe3e3-8abb-40ef-a0ea-93334e6fe94e-build-proxy-ca-bundles\") pod \"smart-gateway-operator-2-build\" (UID: \"492fe3e3-8abb-40ef-a0ea-93334e6fe94e\") " pod="service-telemetry/smart-gateway-operator-2-build" Dec 03 00:25:09 crc kubenswrapper[4953]: I1203 00:25:09.994822 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/492fe3e3-8abb-40ef-a0ea-93334e6fe94e-buildcachedir\") pod \"smart-gateway-operator-2-build\" (UID: \"492fe3e3-8abb-40ef-a0ea-93334e6fe94e\") " pod="service-telemetry/smart-gateway-operator-2-build" Dec 03 00:25:09 crc kubenswrapper[4953]: I1203 00:25:09.996107 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/492fe3e3-8abb-40ef-a0ea-93334e6fe94e-build-blob-cache\") pod \"smart-gateway-operator-2-build\" (UID: \"492fe3e3-8abb-40ef-a0ea-93334e6fe94e\") " pod="service-telemetry/smart-gateway-operator-2-build" Dec 03 00:25:09 crc kubenswrapper[4953]: I1203 00:25:09.996191 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/492fe3e3-8abb-40ef-a0ea-93334e6fe94e-node-pullsecrets\") pod \"smart-gateway-operator-2-build\" (UID: \"492fe3e3-8abb-40ef-a0ea-93334e6fe94e\") " pod="service-telemetry/smart-gateway-operator-2-build" Dec 03 00:25:09 crc kubenswrapper[4953]: I1203 00:25:09.996445 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/492fe3e3-8abb-40ef-a0ea-93334e6fe94e-buildworkdir\") pod \"smart-gateway-operator-2-build\" (UID: \"492fe3e3-8abb-40ef-a0ea-93334e6fe94e\") " pod="service-telemetry/smart-gateway-operator-2-build" Dec 03 00:25:09 crc kubenswrapper[4953]: I1203 00:25:09.997332 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/492fe3e3-8abb-40ef-a0ea-93334e6fe94e-build-ca-bundles\") pod \"smart-gateway-operator-2-build\" (UID: \"492fe3e3-8abb-40ef-a0ea-93334e6fe94e\") " pod="service-telemetry/smart-gateway-operator-2-build" Dec 03 00:25:09 crc kubenswrapper[4953]: I1203 00:25:09.997642 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/492fe3e3-8abb-40ef-a0ea-93334e6fe94e-container-storage-run\") pod \"smart-gateway-operator-2-build\" (UID: \"492fe3e3-8abb-40ef-a0ea-93334e6fe94e\") " pod="service-telemetry/smart-gateway-operator-2-build" Dec 03 00:25:10 crc kubenswrapper[4953]: I1203 00:25:10.004801 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"builder-dockercfg-9prwd-pull\" (UniqueName: \"kubernetes.io/secret/492fe3e3-8abb-40ef-a0ea-93334e6fe94e-builder-dockercfg-9prwd-pull\") pod \"smart-gateway-operator-2-build\" (UID: \"492fe3e3-8abb-40ef-a0ea-93334e6fe94e\") " pod="service-telemetry/smart-gateway-operator-2-build" Dec 03 00:25:10 crc kubenswrapper[4953]: I1203 00:25:10.005148 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"builder-dockercfg-9prwd-push\" (UniqueName: \"kubernetes.io/secret/492fe3e3-8abb-40ef-a0ea-93334e6fe94e-builder-dockercfg-9prwd-push\") pod \"smart-gateway-operator-2-build\" (UID: \"492fe3e3-8abb-40ef-a0ea-93334e6fe94e\") " pod="service-telemetry/smart-gateway-operator-2-build" Dec 03 00:25:10 crc kubenswrapper[4953]: I1203 00:25:10.013214 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sv78b\" (UniqueName: \"kubernetes.io/projected/492fe3e3-8abb-40ef-a0ea-93334e6fe94e-kube-api-access-sv78b\") pod \"smart-gateway-operator-2-build\" (UID: \"492fe3e3-8abb-40ef-a0ea-93334e6fe94e\") " pod="service-telemetry/smart-gateway-operator-2-build" Dec 03 00:25:10 crc kubenswrapper[4953]: I1203 00:25:10.087513 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/smart-gateway-operator-2-build" Dec 03 00:25:10 crc kubenswrapper[4953]: I1203 00:25:10.215712 4953 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_smart-gateway-operator-1-build_2f5ad768-60c4-430b-85c1-90a0f8cddc62/docker-build/0.log" Dec 03 00:25:10 crc kubenswrapper[4953]: I1203 00:25:10.220696 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/smart-gateway-operator-1-build" event={"ID":"2f5ad768-60c4-430b-85c1-90a0f8cddc62","Type":"ContainerDied","Data":"5c4c8d26b24eaec1c9dbd304158cd9e86dd754b45d5e5051f856f03f550c7873"} Dec 03 00:25:10 crc kubenswrapper[4953]: I1203 00:25:10.220788 4953 scope.go:117] "RemoveContainer" containerID="f81776546e6a27a3ab999ecd53085aa3aaf11a750cdfc71dce39d55017cb87b3" Dec 03 00:25:10 crc kubenswrapper[4953]: I1203 00:25:10.220855 4953 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/smart-gateway-operator-1-build" Dec 03 00:25:10 crc kubenswrapper[4953]: I1203 00:25:10.279527 4953 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["service-telemetry/smart-gateway-operator-1-build"] Dec 03 00:25:10 crc kubenswrapper[4953]: I1203 00:25:10.287032 4953 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["service-telemetry/smart-gateway-operator-1-build"] Dec 03 00:25:10 crc kubenswrapper[4953]: I1203 00:25:10.334917 4953 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/smart-gateway-operator-2-build"] Dec 03 00:25:10 crc kubenswrapper[4953]: I1203 00:25:10.352395 4953 scope.go:117] "RemoveContainer" containerID="24ef1cb9f22e5dba721a2f96e5ec39c2559422a67ff70d99e0fff3724af55b5a" Dec 03 00:25:11 crc kubenswrapper[4953]: I1203 00:25:11.229385 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/smart-gateway-operator-2-build" event={"ID":"492fe3e3-8abb-40ef-a0ea-93334e6fe94e","Type":"ContainerStarted","Data":"b0d5158e23b41f86cdcc959621104344c0d62195b594f38e65fea37db9efaeef"} Dec 03 00:25:11 crc kubenswrapper[4953]: I1203 00:25:11.230054 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/smart-gateway-operator-2-build" event={"ID":"492fe3e3-8abb-40ef-a0ea-93334e6fe94e","Type":"ContainerStarted","Data":"dd1aa5598e70b0ab9d4a109b2f89354388821af80021747affd8f645971f1704"} Dec 03 00:25:11 crc kubenswrapper[4953]: I1203 00:25:11.366255 4953 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2f5ad768-60c4-430b-85c1-90a0f8cddc62" path="/var/lib/kubelet/pods/2f5ad768-60c4-430b-85c1-90a0f8cddc62/volumes" Dec 03 00:25:12 crc kubenswrapper[4953]: I1203 00:25:12.254166 4953 generic.go:334] "Generic (PLEG): container finished" podID="492fe3e3-8abb-40ef-a0ea-93334e6fe94e" containerID="b0d5158e23b41f86cdcc959621104344c0d62195b594f38e65fea37db9efaeef" exitCode=0 Dec 03 00:25:12 crc kubenswrapper[4953]: I1203 00:25:12.254238 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/smart-gateway-operator-2-build" event={"ID":"492fe3e3-8abb-40ef-a0ea-93334e6fe94e","Type":"ContainerDied","Data":"b0d5158e23b41f86cdcc959621104344c0d62195b594f38e65fea37db9efaeef"} Dec 03 00:25:13 crc kubenswrapper[4953]: I1203 00:25:13.283396 4953 generic.go:334] "Generic (PLEG): container finished" podID="492fe3e3-8abb-40ef-a0ea-93334e6fe94e" containerID="69a7e00b9c52e5d8d07ad70649caa0072d70d82628fd4523bcd4caa21114fba4" exitCode=0 Dec 03 00:25:13 crc kubenswrapper[4953]: I1203 00:25:13.283485 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/smart-gateway-operator-2-build" event={"ID":"492fe3e3-8abb-40ef-a0ea-93334e6fe94e","Type":"ContainerDied","Data":"69a7e00b9c52e5d8d07ad70649caa0072d70d82628fd4523bcd4caa21114fba4"} Dec 03 00:25:13 crc kubenswrapper[4953]: I1203 00:25:13.325213 4953 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_smart-gateway-operator-2-build_492fe3e3-8abb-40ef-a0ea-93334e6fe94e/manage-dockerfile/0.log" Dec 03 00:25:14 crc kubenswrapper[4953]: I1203 00:25:14.297010 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/smart-gateway-operator-2-build" event={"ID":"492fe3e3-8abb-40ef-a0ea-93334e6fe94e","Type":"ContainerStarted","Data":"82de8276747e537a193fe0f3dd290f42d4673905b215b9980e5633a6ff009142"} Dec 03 00:25:14 crc kubenswrapper[4953]: I1203 00:25:14.337637 4953 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="service-telemetry/smart-gateway-operator-2-build" podStartSLOduration=5.337605087 podStartE2EDuration="5.337605087s" podCreationTimestamp="2025-12-03 00:25:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 00:25:14.332755036 +0000 UTC m=+1103.317035829" watchObservedRunningTime="2025-12-03 00:25:14.337605087 +0000 UTC m=+1103.321885870" Dec 03 00:25:52 crc kubenswrapper[4953]: I1203 00:25:52.602582 4953 trace.go:236] Trace[2137907527]: "Calculate volume metrics of container-storage-root for pod service-telemetry/smart-gateway-operator-2-build" (03-Dec-2025 00:25:51.327) (total time: 1275ms): Dec 03 00:25:52 crc kubenswrapper[4953]: Trace[2137907527]: [1.275137076s] [1.275137076s] END Dec 03 00:26:18 crc kubenswrapper[4953]: I1203 00:26:18.945165 4953 patch_prober.go:28] interesting pod/machine-config-daemon-6dfbn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 00:26:18 crc kubenswrapper[4953]: I1203 00:26:18.946157 4953 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-6dfbn" podUID="53ffb1e9-9a2e-4869-80d7-776f4aa7ef8b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 00:26:44 crc kubenswrapper[4953]: I1203 00:26:44.947699 4953 generic.go:334] "Generic (PLEG): container finished" podID="492fe3e3-8abb-40ef-a0ea-93334e6fe94e" containerID="82de8276747e537a193fe0f3dd290f42d4673905b215b9980e5633a6ff009142" exitCode=0 Dec 03 00:26:44 crc kubenswrapper[4953]: I1203 00:26:44.947818 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/smart-gateway-operator-2-build" event={"ID":"492fe3e3-8abb-40ef-a0ea-93334e6fe94e","Type":"ContainerDied","Data":"82de8276747e537a193fe0f3dd290f42d4673905b215b9980e5633a6ff009142"} Dec 03 00:26:46 crc kubenswrapper[4953]: I1203 00:26:46.332236 4953 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/smart-gateway-operator-2-build" Dec 03 00:26:46 crc kubenswrapper[4953]: I1203 00:26:46.392526 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/492fe3e3-8abb-40ef-a0ea-93334e6fe94e-container-storage-root\") pod \"492fe3e3-8abb-40ef-a0ea-93334e6fe94e\" (UID: \"492fe3e3-8abb-40ef-a0ea-93334e6fe94e\") " Dec 03 00:26:46 crc kubenswrapper[4953]: I1203 00:26:46.392809 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/492fe3e3-8abb-40ef-a0ea-93334e6fe94e-build-system-configs\") pod \"492fe3e3-8abb-40ef-a0ea-93334e6fe94e\" (UID: \"492fe3e3-8abb-40ef-a0ea-93334e6fe94e\") " Dec 03 00:26:46 crc kubenswrapper[4953]: I1203 00:26:46.392838 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/492fe3e3-8abb-40ef-a0ea-93334e6fe94e-node-pullsecrets\") pod \"492fe3e3-8abb-40ef-a0ea-93334e6fe94e\" (UID: \"492fe3e3-8abb-40ef-a0ea-93334e6fe94e\") " Dec 03 00:26:46 crc kubenswrapper[4953]: I1203 00:26:46.392871 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sv78b\" (UniqueName: \"kubernetes.io/projected/492fe3e3-8abb-40ef-a0ea-93334e6fe94e-kube-api-access-sv78b\") pod \"492fe3e3-8abb-40ef-a0ea-93334e6fe94e\" (UID: \"492fe3e3-8abb-40ef-a0ea-93334e6fe94e\") " Dec 03 00:26:46 crc kubenswrapper[4953]: I1203 00:26:46.392899 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/492fe3e3-8abb-40ef-a0ea-93334e6fe94e-build-ca-bundles\") pod \"492fe3e3-8abb-40ef-a0ea-93334e6fe94e\" (UID: \"492fe3e3-8abb-40ef-a0ea-93334e6fe94e\") " Dec 03 00:26:46 crc kubenswrapper[4953]: I1203 00:26:46.392920 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/492fe3e3-8abb-40ef-a0ea-93334e6fe94e-buildworkdir\") pod \"492fe3e3-8abb-40ef-a0ea-93334e6fe94e\" (UID: \"492fe3e3-8abb-40ef-a0ea-93334e6fe94e\") " Dec 03 00:26:46 crc kubenswrapper[4953]: I1203 00:26:46.392943 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"builder-dockercfg-9prwd-pull\" (UniqueName: \"kubernetes.io/secret/492fe3e3-8abb-40ef-a0ea-93334e6fe94e-builder-dockercfg-9prwd-pull\") pod \"492fe3e3-8abb-40ef-a0ea-93334e6fe94e\" (UID: \"492fe3e3-8abb-40ef-a0ea-93334e6fe94e\") " Dec 03 00:26:46 crc kubenswrapper[4953]: I1203 00:26:46.392988 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"builder-dockercfg-9prwd-push\" (UniqueName: \"kubernetes.io/secret/492fe3e3-8abb-40ef-a0ea-93334e6fe94e-builder-dockercfg-9prwd-push\") pod \"492fe3e3-8abb-40ef-a0ea-93334e6fe94e\" (UID: \"492fe3e3-8abb-40ef-a0ea-93334e6fe94e\") " Dec 03 00:26:46 crc kubenswrapper[4953]: I1203 00:26:46.393018 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/492fe3e3-8abb-40ef-a0ea-93334e6fe94e-container-storage-run\") pod \"492fe3e3-8abb-40ef-a0ea-93334e6fe94e\" (UID: \"492fe3e3-8abb-40ef-a0ea-93334e6fe94e\") " Dec 03 00:26:46 crc kubenswrapper[4953]: I1203 00:26:46.393060 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/492fe3e3-8abb-40ef-a0ea-93334e6fe94e-build-proxy-ca-bundles\") pod \"492fe3e3-8abb-40ef-a0ea-93334e6fe94e\" (UID: \"492fe3e3-8abb-40ef-a0ea-93334e6fe94e\") " Dec 03 00:26:46 crc kubenswrapper[4953]: I1203 00:26:46.393081 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/492fe3e3-8abb-40ef-a0ea-93334e6fe94e-build-blob-cache\") pod \"492fe3e3-8abb-40ef-a0ea-93334e6fe94e\" (UID: \"492fe3e3-8abb-40ef-a0ea-93334e6fe94e\") " Dec 03 00:26:46 crc kubenswrapper[4953]: I1203 00:26:46.393099 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/492fe3e3-8abb-40ef-a0ea-93334e6fe94e-buildcachedir\") pod \"492fe3e3-8abb-40ef-a0ea-93334e6fe94e\" (UID: \"492fe3e3-8abb-40ef-a0ea-93334e6fe94e\") " Dec 03 00:26:46 crc kubenswrapper[4953]: I1203 00:26:46.393305 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/492fe3e3-8abb-40ef-a0ea-93334e6fe94e-buildcachedir" (OuterVolumeSpecName: "buildcachedir") pod "492fe3e3-8abb-40ef-a0ea-93334e6fe94e" (UID: "492fe3e3-8abb-40ef-a0ea-93334e6fe94e"). InnerVolumeSpecName "buildcachedir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 00:26:46 crc kubenswrapper[4953]: I1203 00:26:46.395954 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/492fe3e3-8abb-40ef-a0ea-93334e6fe94e-container-storage-run" (OuterVolumeSpecName: "container-storage-run") pod "492fe3e3-8abb-40ef-a0ea-93334e6fe94e" (UID: "492fe3e3-8abb-40ef-a0ea-93334e6fe94e"). InnerVolumeSpecName "container-storage-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 00:26:46 crc kubenswrapper[4953]: I1203 00:26:46.398422 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/492fe3e3-8abb-40ef-a0ea-93334e6fe94e-build-ca-bundles" (OuterVolumeSpecName: "build-ca-bundles") pod "492fe3e3-8abb-40ef-a0ea-93334e6fe94e" (UID: "492fe3e3-8abb-40ef-a0ea-93334e6fe94e"). InnerVolumeSpecName "build-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:26:46 crc kubenswrapper[4953]: I1203 00:26:46.399042 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/492fe3e3-8abb-40ef-a0ea-93334e6fe94e-build-system-configs" (OuterVolumeSpecName: "build-system-configs") pod "492fe3e3-8abb-40ef-a0ea-93334e6fe94e" (UID: "492fe3e3-8abb-40ef-a0ea-93334e6fe94e"). InnerVolumeSpecName "build-system-configs". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:26:46 crc kubenswrapper[4953]: I1203 00:26:46.399250 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/492fe3e3-8abb-40ef-a0ea-93334e6fe94e-node-pullsecrets" (OuterVolumeSpecName: "node-pullsecrets") pod "492fe3e3-8abb-40ef-a0ea-93334e6fe94e" (UID: "492fe3e3-8abb-40ef-a0ea-93334e6fe94e"). InnerVolumeSpecName "node-pullsecrets". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 00:26:46 crc kubenswrapper[4953]: I1203 00:26:46.399939 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/492fe3e3-8abb-40ef-a0ea-93334e6fe94e-build-proxy-ca-bundles" (OuterVolumeSpecName: "build-proxy-ca-bundles") pod "492fe3e3-8abb-40ef-a0ea-93334e6fe94e" (UID: "492fe3e3-8abb-40ef-a0ea-93334e6fe94e"). InnerVolumeSpecName "build-proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:26:46 crc kubenswrapper[4953]: I1203 00:26:46.402044 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/492fe3e3-8abb-40ef-a0ea-93334e6fe94e-builder-dockercfg-9prwd-push" (OuterVolumeSpecName: "builder-dockercfg-9prwd-push") pod "492fe3e3-8abb-40ef-a0ea-93334e6fe94e" (UID: "492fe3e3-8abb-40ef-a0ea-93334e6fe94e"). InnerVolumeSpecName "builder-dockercfg-9prwd-push". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:26:46 crc kubenswrapper[4953]: I1203 00:26:46.408200 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/492fe3e3-8abb-40ef-a0ea-93334e6fe94e-builder-dockercfg-9prwd-pull" (OuterVolumeSpecName: "builder-dockercfg-9prwd-pull") pod "492fe3e3-8abb-40ef-a0ea-93334e6fe94e" (UID: "492fe3e3-8abb-40ef-a0ea-93334e6fe94e"). InnerVolumeSpecName "builder-dockercfg-9prwd-pull". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:26:46 crc kubenswrapper[4953]: I1203 00:26:46.408705 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/492fe3e3-8abb-40ef-a0ea-93334e6fe94e-kube-api-access-sv78b" (OuterVolumeSpecName: "kube-api-access-sv78b") pod "492fe3e3-8abb-40ef-a0ea-93334e6fe94e" (UID: "492fe3e3-8abb-40ef-a0ea-93334e6fe94e"). InnerVolumeSpecName "kube-api-access-sv78b". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:26:46 crc kubenswrapper[4953]: I1203 00:26:46.414021 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/492fe3e3-8abb-40ef-a0ea-93334e6fe94e-buildworkdir" (OuterVolumeSpecName: "buildworkdir") pod "492fe3e3-8abb-40ef-a0ea-93334e6fe94e" (UID: "492fe3e3-8abb-40ef-a0ea-93334e6fe94e"). InnerVolumeSpecName "buildworkdir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 00:26:46 crc kubenswrapper[4953]: I1203 00:26:46.494174 4953 reconciler_common.go:293] "Volume detached for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/492fe3e3-8abb-40ef-a0ea-93334e6fe94e-build-system-configs\") on node \"crc\" DevicePath \"\"" Dec 03 00:26:46 crc kubenswrapper[4953]: I1203 00:26:46.494207 4953 reconciler_common.go:293] "Volume detached for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/492fe3e3-8abb-40ef-a0ea-93334e6fe94e-node-pullsecrets\") on node \"crc\" DevicePath \"\"" Dec 03 00:26:46 crc kubenswrapper[4953]: I1203 00:26:46.494217 4953 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sv78b\" (UniqueName: \"kubernetes.io/projected/492fe3e3-8abb-40ef-a0ea-93334e6fe94e-kube-api-access-sv78b\") on node \"crc\" DevicePath \"\"" Dec 03 00:26:46 crc kubenswrapper[4953]: I1203 00:26:46.494224 4953 reconciler_common.go:293] "Volume detached for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/492fe3e3-8abb-40ef-a0ea-93334e6fe94e-build-ca-bundles\") on node \"crc\" DevicePath \"\"" Dec 03 00:26:46 crc kubenswrapper[4953]: I1203 00:26:46.494235 4953 reconciler_common.go:293] "Volume detached for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/492fe3e3-8abb-40ef-a0ea-93334e6fe94e-buildworkdir\") on node \"crc\" DevicePath \"\"" Dec 03 00:26:46 crc kubenswrapper[4953]: I1203 00:26:46.494243 4953 reconciler_common.go:293] "Volume detached for volume \"builder-dockercfg-9prwd-pull\" (UniqueName: \"kubernetes.io/secret/492fe3e3-8abb-40ef-a0ea-93334e6fe94e-builder-dockercfg-9prwd-pull\") on node \"crc\" DevicePath \"\"" Dec 03 00:26:46 crc kubenswrapper[4953]: I1203 00:26:46.494253 4953 reconciler_common.go:293] "Volume detached for volume \"builder-dockercfg-9prwd-push\" (UniqueName: \"kubernetes.io/secret/492fe3e3-8abb-40ef-a0ea-93334e6fe94e-builder-dockercfg-9prwd-push\") on node \"crc\" DevicePath \"\"" Dec 03 00:26:46 crc kubenswrapper[4953]: I1203 00:26:46.494261 4953 reconciler_common.go:293] "Volume detached for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/492fe3e3-8abb-40ef-a0ea-93334e6fe94e-container-storage-run\") on node \"crc\" DevicePath \"\"" Dec 03 00:26:46 crc kubenswrapper[4953]: I1203 00:26:46.494270 4953 reconciler_common.go:293] "Volume detached for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/492fe3e3-8abb-40ef-a0ea-93334e6fe94e-build-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Dec 03 00:26:46 crc kubenswrapper[4953]: I1203 00:26:46.494278 4953 reconciler_common.go:293] "Volume detached for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/492fe3e3-8abb-40ef-a0ea-93334e6fe94e-buildcachedir\") on node \"crc\" DevicePath \"\"" Dec 03 00:26:46 crc kubenswrapper[4953]: I1203 00:26:46.617231 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/492fe3e3-8abb-40ef-a0ea-93334e6fe94e-build-blob-cache" (OuterVolumeSpecName: "build-blob-cache") pod "492fe3e3-8abb-40ef-a0ea-93334e6fe94e" (UID: "492fe3e3-8abb-40ef-a0ea-93334e6fe94e"). InnerVolumeSpecName "build-blob-cache". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 00:26:46 crc kubenswrapper[4953]: I1203 00:26:46.697316 4953 reconciler_common.go:293] "Volume detached for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/492fe3e3-8abb-40ef-a0ea-93334e6fe94e-build-blob-cache\") on node \"crc\" DevicePath \"\"" Dec 03 00:26:46 crc kubenswrapper[4953]: I1203 00:26:46.973625 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/smart-gateway-operator-2-build" event={"ID":"492fe3e3-8abb-40ef-a0ea-93334e6fe94e","Type":"ContainerDied","Data":"dd1aa5598e70b0ab9d4a109b2f89354388821af80021747affd8f645971f1704"} Dec 03 00:26:46 crc kubenswrapper[4953]: I1203 00:26:46.973685 4953 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="dd1aa5598e70b0ab9d4a109b2f89354388821af80021747affd8f645971f1704" Dec 03 00:26:46 crc kubenswrapper[4953]: I1203 00:26:46.973739 4953 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/smart-gateway-operator-2-build" Dec 03 00:26:48 crc kubenswrapper[4953]: I1203 00:26:48.944303 4953 patch_prober.go:28] interesting pod/machine-config-daemon-6dfbn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 00:26:48 crc kubenswrapper[4953]: I1203 00:26:48.944376 4953 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-6dfbn" podUID="53ffb1e9-9a2e-4869-80d7-776f4aa7ef8b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 00:26:49 crc kubenswrapper[4953]: I1203 00:26:49.060167 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/492fe3e3-8abb-40ef-a0ea-93334e6fe94e-container-storage-root" (OuterVolumeSpecName: "container-storage-root") pod "492fe3e3-8abb-40ef-a0ea-93334e6fe94e" (UID: "492fe3e3-8abb-40ef-a0ea-93334e6fe94e"). InnerVolumeSpecName "container-storage-root". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 00:26:49 crc kubenswrapper[4953]: I1203 00:26:49.133713 4953 reconciler_common.go:293] "Volume detached for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/492fe3e3-8abb-40ef-a0ea-93334e6fe94e-container-storage-root\") on node \"crc\" DevicePath \"\"" Dec 03 00:26:50 crc kubenswrapper[4953]: I1203 00:26:50.873240 4953 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["service-telemetry/sg-core-1-build"] Dec 03 00:26:50 crc kubenswrapper[4953]: E1203 00:26:50.873467 4953 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="492fe3e3-8abb-40ef-a0ea-93334e6fe94e" containerName="git-clone" Dec 03 00:26:50 crc kubenswrapper[4953]: I1203 00:26:50.873480 4953 state_mem.go:107] "Deleted CPUSet assignment" podUID="492fe3e3-8abb-40ef-a0ea-93334e6fe94e" containerName="git-clone" Dec 03 00:26:50 crc kubenswrapper[4953]: E1203 00:26:50.873487 4953 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="492fe3e3-8abb-40ef-a0ea-93334e6fe94e" containerName="docker-build" Dec 03 00:26:50 crc kubenswrapper[4953]: I1203 00:26:50.873493 4953 state_mem.go:107] "Deleted CPUSet assignment" podUID="492fe3e3-8abb-40ef-a0ea-93334e6fe94e" containerName="docker-build" Dec 03 00:26:50 crc kubenswrapper[4953]: E1203 00:26:50.873509 4953 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="492fe3e3-8abb-40ef-a0ea-93334e6fe94e" containerName="manage-dockerfile" Dec 03 00:26:50 crc kubenswrapper[4953]: I1203 00:26:50.873516 4953 state_mem.go:107] "Deleted CPUSet assignment" podUID="492fe3e3-8abb-40ef-a0ea-93334e6fe94e" containerName="manage-dockerfile" Dec 03 00:26:50 crc kubenswrapper[4953]: I1203 00:26:50.873603 4953 memory_manager.go:354] "RemoveStaleState removing state" podUID="492fe3e3-8abb-40ef-a0ea-93334e6fe94e" containerName="docker-build" Dec 03 00:26:50 crc kubenswrapper[4953]: I1203 00:26:50.874158 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/sg-core-1-build" Dec 03 00:26:50 crc kubenswrapper[4953]: I1203 00:26:50.876542 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"sg-core-1-sys-config" Dec 03 00:26:50 crc kubenswrapper[4953]: I1203 00:26:50.876757 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"sg-core-1-global-ca" Dec 03 00:26:50 crc kubenswrapper[4953]: I1203 00:26:50.876885 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"sg-core-1-ca" Dec 03 00:26:50 crc kubenswrapper[4953]: I1203 00:26:50.877083 4953 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"builder-dockercfg-9prwd" Dec 03 00:26:50 crc kubenswrapper[4953]: I1203 00:26:50.889946 4953 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/sg-core-1-build"] Dec 03 00:26:51 crc kubenswrapper[4953]: I1203 00:26:51.029780 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/5cd4ce9d-4089-4482-8910-83128307799d-container-storage-root\") pod \"sg-core-1-build\" (UID: \"5cd4ce9d-4089-4482-8910-83128307799d\") " pod="service-telemetry/sg-core-1-build" Dec 03 00:26:51 crc kubenswrapper[4953]: I1203 00:26:51.029828 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7grbf\" (UniqueName: \"kubernetes.io/projected/5cd4ce9d-4089-4482-8910-83128307799d-kube-api-access-7grbf\") pod \"sg-core-1-build\" (UID: \"5cd4ce9d-4089-4482-8910-83128307799d\") " pod="service-telemetry/sg-core-1-build" Dec 03 00:26:51 crc kubenswrapper[4953]: I1203 00:26:51.029847 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/5cd4ce9d-4089-4482-8910-83128307799d-node-pullsecrets\") pod \"sg-core-1-build\" (UID: \"5cd4ce9d-4089-4482-8910-83128307799d\") " pod="service-telemetry/sg-core-1-build" Dec 03 00:26:51 crc kubenswrapper[4953]: I1203 00:26:51.029866 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/5cd4ce9d-4089-4482-8910-83128307799d-container-storage-run\") pod \"sg-core-1-build\" (UID: \"5cd4ce9d-4089-4482-8910-83128307799d\") " pod="service-telemetry/sg-core-1-build" Dec 03 00:26:51 crc kubenswrapper[4953]: I1203 00:26:51.029895 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/5cd4ce9d-4089-4482-8910-83128307799d-buildworkdir\") pod \"sg-core-1-build\" (UID: \"5cd4ce9d-4089-4482-8910-83128307799d\") " pod="service-telemetry/sg-core-1-build" Dec 03 00:26:51 crc kubenswrapper[4953]: I1203 00:26:51.029917 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/5cd4ce9d-4089-4482-8910-83128307799d-build-blob-cache\") pod \"sg-core-1-build\" (UID: \"5cd4ce9d-4089-4482-8910-83128307799d\") " pod="service-telemetry/sg-core-1-build" Dec 03 00:26:51 crc kubenswrapper[4953]: I1203 00:26:51.029935 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/5cd4ce9d-4089-4482-8910-83128307799d-build-system-configs\") pod \"sg-core-1-build\" (UID: \"5cd4ce9d-4089-4482-8910-83128307799d\") " pod="service-telemetry/sg-core-1-build" Dec 03 00:26:51 crc kubenswrapper[4953]: I1203 00:26:51.029950 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/5cd4ce9d-4089-4482-8910-83128307799d-build-ca-bundles\") pod \"sg-core-1-build\" (UID: \"5cd4ce9d-4089-4482-8910-83128307799d\") " pod="service-telemetry/sg-core-1-build" Dec 03 00:26:51 crc kubenswrapper[4953]: I1203 00:26:51.029993 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/5cd4ce9d-4089-4482-8910-83128307799d-buildcachedir\") pod \"sg-core-1-build\" (UID: \"5cd4ce9d-4089-4482-8910-83128307799d\") " pod="service-telemetry/sg-core-1-build" Dec 03 00:26:51 crc kubenswrapper[4953]: I1203 00:26:51.030013 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"builder-dockercfg-9prwd-push\" (UniqueName: \"kubernetes.io/secret/5cd4ce9d-4089-4482-8910-83128307799d-builder-dockercfg-9prwd-push\") pod \"sg-core-1-build\" (UID: \"5cd4ce9d-4089-4482-8910-83128307799d\") " pod="service-telemetry/sg-core-1-build" Dec 03 00:26:51 crc kubenswrapper[4953]: I1203 00:26:51.030033 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"builder-dockercfg-9prwd-pull\" (UniqueName: \"kubernetes.io/secret/5cd4ce9d-4089-4482-8910-83128307799d-builder-dockercfg-9prwd-pull\") pod \"sg-core-1-build\" (UID: \"5cd4ce9d-4089-4482-8910-83128307799d\") " pod="service-telemetry/sg-core-1-build" Dec 03 00:26:51 crc kubenswrapper[4953]: I1203 00:26:51.030047 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/5cd4ce9d-4089-4482-8910-83128307799d-build-proxy-ca-bundles\") pod \"sg-core-1-build\" (UID: \"5cd4ce9d-4089-4482-8910-83128307799d\") " pod="service-telemetry/sg-core-1-build" Dec 03 00:26:51 crc kubenswrapper[4953]: I1203 00:26:51.131271 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/5cd4ce9d-4089-4482-8910-83128307799d-build-blob-cache\") pod \"sg-core-1-build\" (UID: \"5cd4ce9d-4089-4482-8910-83128307799d\") " pod="service-telemetry/sg-core-1-build" Dec 03 00:26:51 crc kubenswrapper[4953]: I1203 00:26:51.131330 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/5cd4ce9d-4089-4482-8910-83128307799d-build-system-configs\") pod \"sg-core-1-build\" (UID: \"5cd4ce9d-4089-4482-8910-83128307799d\") " pod="service-telemetry/sg-core-1-build" Dec 03 00:26:51 crc kubenswrapper[4953]: I1203 00:26:51.131356 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/5cd4ce9d-4089-4482-8910-83128307799d-build-ca-bundles\") pod \"sg-core-1-build\" (UID: \"5cd4ce9d-4089-4482-8910-83128307799d\") " pod="service-telemetry/sg-core-1-build" Dec 03 00:26:51 crc kubenswrapper[4953]: I1203 00:26:51.131401 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/5cd4ce9d-4089-4482-8910-83128307799d-buildcachedir\") pod \"sg-core-1-build\" (UID: \"5cd4ce9d-4089-4482-8910-83128307799d\") " pod="service-telemetry/sg-core-1-build" Dec 03 00:26:51 crc kubenswrapper[4953]: I1203 00:26:51.131431 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"builder-dockercfg-9prwd-push\" (UniqueName: \"kubernetes.io/secret/5cd4ce9d-4089-4482-8910-83128307799d-builder-dockercfg-9prwd-push\") pod \"sg-core-1-build\" (UID: \"5cd4ce9d-4089-4482-8910-83128307799d\") " pod="service-telemetry/sg-core-1-build" Dec 03 00:26:51 crc kubenswrapper[4953]: I1203 00:26:51.131460 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"builder-dockercfg-9prwd-pull\" (UniqueName: \"kubernetes.io/secret/5cd4ce9d-4089-4482-8910-83128307799d-builder-dockercfg-9prwd-pull\") pod \"sg-core-1-build\" (UID: \"5cd4ce9d-4089-4482-8910-83128307799d\") " pod="service-telemetry/sg-core-1-build" Dec 03 00:26:51 crc kubenswrapper[4953]: I1203 00:26:51.131483 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/5cd4ce9d-4089-4482-8910-83128307799d-build-proxy-ca-bundles\") pod \"sg-core-1-build\" (UID: \"5cd4ce9d-4089-4482-8910-83128307799d\") " pod="service-telemetry/sg-core-1-build" Dec 03 00:26:51 crc kubenswrapper[4953]: I1203 00:26:51.131539 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/5cd4ce9d-4089-4482-8910-83128307799d-node-pullsecrets\") pod \"sg-core-1-build\" (UID: \"5cd4ce9d-4089-4482-8910-83128307799d\") " pod="service-telemetry/sg-core-1-build" Dec 03 00:26:51 crc kubenswrapper[4953]: I1203 00:26:51.131560 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/5cd4ce9d-4089-4482-8910-83128307799d-container-storage-root\") pod \"sg-core-1-build\" (UID: \"5cd4ce9d-4089-4482-8910-83128307799d\") " pod="service-telemetry/sg-core-1-build" Dec 03 00:26:51 crc kubenswrapper[4953]: I1203 00:26:51.131582 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7grbf\" (UniqueName: \"kubernetes.io/projected/5cd4ce9d-4089-4482-8910-83128307799d-kube-api-access-7grbf\") pod \"sg-core-1-build\" (UID: \"5cd4ce9d-4089-4482-8910-83128307799d\") " pod="service-telemetry/sg-core-1-build" Dec 03 00:26:51 crc kubenswrapper[4953]: I1203 00:26:51.131616 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/5cd4ce9d-4089-4482-8910-83128307799d-container-storage-run\") pod \"sg-core-1-build\" (UID: \"5cd4ce9d-4089-4482-8910-83128307799d\") " pod="service-telemetry/sg-core-1-build" Dec 03 00:26:51 crc kubenswrapper[4953]: I1203 00:26:51.131651 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/5cd4ce9d-4089-4482-8910-83128307799d-buildworkdir\") pod \"sg-core-1-build\" (UID: \"5cd4ce9d-4089-4482-8910-83128307799d\") " pod="service-telemetry/sg-core-1-build" Dec 03 00:26:51 crc kubenswrapper[4953]: I1203 00:26:51.132144 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/5cd4ce9d-4089-4482-8910-83128307799d-buildworkdir\") pod \"sg-core-1-build\" (UID: \"5cd4ce9d-4089-4482-8910-83128307799d\") " pod="service-telemetry/sg-core-1-build" Dec 03 00:26:51 crc kubenswrapper[4953]: I1203 00:26:51.132385 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/5cd4ce9d-4089-4482-8910-83128307799d-build-blob-cache\") pod \"sg-core-1-build\" (UID: \"5cd4ce9d-4089-4482-8910-83128307799d\") " pod="service-telemetry/sg-core-1-build" Dec 03 00:26:51 crc kubenswrapper[4953]: I1203 00:26:51.132722 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/5cd4ce9d-4089-4482-8910-83128307799d-buildcachedir\") pod \"sg-core-1-build\" (UID: \"5cd4ce9d-4089-4482-8910-83128307799d\") " pod="service-telemetry/sg-core-1-build" Dec 03 00:26:51 crc kubenswrapper[4953]: I1203 00:26:51.132933 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/5cd4ce9d-4089-4482-8910-83128307799d-build-system-configs\") pod \"sg-core-1-build\" (UID: \"5cd4ce9d-4089-4482-8910-83128307799d\") " pod="service-telemetry/sg-core-1-build" Dec 03 00:26:51 crc kubenswrapper[4953]: I1203 00:26:51.133193 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/5cd4ce9d-4089-4482-8910-83128307799d-container-storage-root\") pod \"sg-core-1-build\" (UID: \"5cd4ce9d-4089-4482-8910-83128307799d\") " pod="service-telemetry/sg-core-1-build" Dec 03 00:26:51 crc kubenswrapper[4953]: I1203 00:26:51.133293 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/5cd4ce9d-4089-4482-8910-83128307799d-node-pullsecrets\") pod \"sg-core-1-build\" (UID: \"5cd4ce9d-4089-4482-8910-83128307799d\") " pod="service-telemetry/sg-core-1-build" Dec 03 00:26:51 crc kubenswrapper[4953]: I1203 00:26:51.133409 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/5cd4ce9d-4089-4482-8910-83128307799d-build-ca-bundles\") pod \"sg-core-1-build\" (UID: \"5cd4ce9d-4089-4482-8910-83128307799d\") " pod="service-telemetry/sg-core-1-build" Dec 03 00:26:51 crc kubenswrapper[4953]: I1203 00:26:51.133521 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/5cd4ce9d-4089-4482-8910-83128307799d-container-storage-run\") pod \"sg-core-1-build\" (UID: \"5cd4ce9d-4089-4482-8910-83128307799d\") " pod="service-telemetry/sg-core-1-build" Dec 03 00:26:51 crc kubenswrapper[4953]: I1203 00:26:51.134489 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/5cd4ce9d-4089-4482-8910-83128307799d-build-proxy-ca-bundles\") pod \"sg-core-1-build\" (UID: \"5cd4ce9d-4089-4482-8910-83128307799d\") " pod="service-telemetry/sg-core-1-build" Dec 03 00:26:51 crc kubenswrapper[4953]: I1203 00:26:51.140410 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"builder-dockercfg-9prwd-pull\" (UniqueName: \"kubernetes.io/secret/5cd4ce9d-4089-4482-8910-83128307799d-builder-dockercfg-9prwd-pull\") pod \"sg-core-1-build\" (UID: \"5cd4ce9d-4089-4482-8910-83128307799d\") " pod="service-telemetry/sg-core-1-build" Dec 03 00:26:51 crc kubenswrapper[4953]: I1203 00:26:51.143435 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"builder-dockercfg-9prwd-push\" (UniqueName: \"kubernetes.io/secret/5cd4ce9d-4089-4482-8910-83128307799d-builder-dockercfg-9prwd-push\") pod \"sg-core-1-build\" (UID: \"5cd4ce9d-4089-4482-8910-83128307799d\") " pod="service-telemetry/sg-core-1-build" Dec 03 00:26:51 crc kubenswrapper[4953]: I1203 00:26:51.149159 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7grbf\" (UniqueName: \"kubernetes.io/projected/5cd4ce9d-4089-4482-8910-83128307799d-kube-api-access-7grbf\") pod \"sg-core-1-build\" (UID: \"5cd4ce9d-4089-4482-8910-83128307799d\") " pod="service-telemetry/sg-core-1-build" Dec 03 00:26:51 crc kubenswrapper[4953]: I1203 00:26:51.229798 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/sg-core-1-build" Dec 03 00:26:51 crc kubenswrapper[4953]: I1203 00:26:51.460023 4953 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/sg-core-1-build"] Dec 03 00:26:52 crc kubenswrapper[4953]: I1203 00:26:52.021231 4953 generic.go:334] "Generic (PLEG): container finished" podID="5cd4ce9d-4089-4482-8910-83128307799d" containerID="b11cc40cc1f04c662d2d8e4e6cb6ca1448e0bd6a7c96c80f6b0ca1e0b74f3149" exitCode=0 Dec 03 00:26:52 crc kubenswrapper[4953]: I1203 00:26:52.021372 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/sg-core-1-build" event={"ID":"5cd4ce9d-4089-4482-8910-83128307799d","Type":"ContainerDied","Data":"b11cc40cc1f04c662d2d8e4e6cb6ca1448e0bd6a7c96c80f6b0ca1e0b74f3149"} Dec 03 00:26:52 crc kubenswrapper[4953]: I1203 00:26:52.021566 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/sg-core-1-build" event={"ID":"5cd4ce9d-4089-4482-8910-83128307799d","Type":"ContainerStarted","Data":"7fcf976e4389cb7ec43e26d5fabe6f6c3da8c70dfd1db50fe11c00a5cf9d3736"} Dec 03 00:26:53 crc kubenswrapper[4953]: I1203 00:26:53.031050 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/sg-core-1-build" event={"ID":"5cd4ce9d-4089-4482-8910-83128307799d","Type":"ContainerStarted","Data":"2d1f0cede27b9aa51a9df95b41c323e6e9330bd930a5170a0a3d629e5d2c9e36"} Dec 03 00:26:53 crc kubenswrapper[4953]: I1203 00:26:53.058061 4953 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="service-telemetry/sg-core-1-build" podStartSLOduration=3.057946179 podStartE2EDuration="3.057946179s" podCreationTimestamp="2025-12-03 00:26:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 00:26:53.056575625 +0000 UTC m=+1202.040856408" watchObservedRunningTime="2025-12-03 00:26:53.057946179 +0000 UTC m=+1202.042227002" Dec 03 00:27:01 crc kubenswrapper[4953]: I1203 00:27:01.303672 4953 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["service-telemetry/sg-core-1-build"] Dec 03 00:27:01 crc kubenswrapper[4953]: I1203 00:27:01.304460 4953 kuberuntime_container.go:808] "Killing container with a grace period" pod="service-telemetry/sg-core-1-build" podUID="5cd4ce9d-4089-4482-8910-83128307799d" containerName="docker-build" containerID="cri-o://2d1f0cede27b9aa51a9df95b41c323e6e9330bd930a5170a0a3d629e5d2c9e36" gracePeriod=30 Dec 03 00:27:01 crc kubenswrapper[4953]: I1203 00:27:01.766206 4953 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_sg-core-1-build_5cd4ce9d-4089-4482-8910-83128307799d/docker-build/0.log" Dec 03 00:27:01 crc kubenswrapper[4953]: I1203 00:27:01.767363 4953 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/sg-core-1-build" Dec 03 00:27:01 crc kubenswrapper[4953]: I1203 00:27:01.894493 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/5cd4ce9d-4089-4482-8910-83128307799d-container-storage-run\") pod \"5cd4ce9d-4089-4482-8910-83128307799d\" (UID: \"5cd4ce9d-4089-4482-8910-83128307799d\") " Dec 03 00:27:01 crc kubenswrapper[4953]: I1203 00:27:01.894692 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7grbf\" (UniqueName: \"kubernetes.io/projected/5cd4ce9d-4089-4482-8910-83128307799d-kube-api-access-7grbf\") pod \"5cd4ce9d-4089-4482-8910-83128307799d\" (UID: \"5cd4ce9d-4089-4482-8910-83128307799d\") " Dec 03 00:27:01 crc kubenswrapper[4953]: I1203 00:27:01.894795 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"builder-dockercfg-9prwd-pull\" (UniqueName: \"kubernetes.io/secret/5cd4ce9d-4089-4482-8910-83128307799d-builder-dockercfg-9prwd-pull\") pod \"5cd4ce9d-4089-4482-8910-83128307799d\" (UID: \"5cd4ce9d-4089-4482-8910-83128307799d\") " Dec 03 00:27:01 crc kubenswrapper[4953]: I1203 00:27:01.895892 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5cd4ce9d-4089-4482-8910-83128307799d-container-storage-run" (OuterVolumeSpecName: "container-storage-run") pod "5cd4ce9d-4089-4482-8910-83128307799d" (UID: "5cd4ce9d-4089-4482-8910-83128307799d"). InnerVolumeSpecName "container-storage-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 00:27:01 crc kubenswrapper[4953]: I1203 00:27:01.896067 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"builder-dockercfg-9prwd-push\" (UniqueName: \"kubernetes.io/secret/5cd4ce9d-4089-4482-8910-83128307799d-builder-dockercfg-9prwd-push\") pod \"5cd4ce9d-4089-4482-8910-83128307799d\" (UID: \"5cd4ce9d-4089-4482-8910-83128307799d\") " Dec 03 00:27:01 crc kubenswrapper[4953]: I1203 00:27:01.896152 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/5cd4ce9d-4089-4482-8910-83128307799d-container-storage-root\") pod \"5cd4ce9d-4089-4482-8910-83128307799d\" (UID: \"5cd4ce9d-4089-4482-8910-83128307799d\") " Dec 03 00:27:01 crc kubenswrapper[4953]: I1203 00:27:01.896210 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/5cd4ce9d-4089-4482-8910-83128307799d-build-proxy-ca-bundles\") pod \"5cd4ce9d-4089-4482-8910-83128307799d\" (UID: \"5cd4ce9d-4089-4482-8910-83128307799d\") " Dec 03 00:27:01 crc kubenswrapper[4953]: I1203 00:27:01.896266 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/5cd4ce9d-4089-4482-8910-83128307799d-node-pullsecrets\") pod \"5cd4ce9d-4089-4482-8910-83128307799d\" (UID: \"5cd4ce9d-4089-4482-8910-83128307799d\") " Dec 03 00:27:01 crc kubenswrapper[4953]: I1203 00:27:01.896363 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/5cd4ce9d-4089-4482-8910-83128307799d-build-system-configs\") pod \"5cd4ce9d-4089-4482-8910-83128307799d\" (UID: \"5cd4ce9d-4089-4482-8910-83128307799d\") " Dec 03 00:27:01 crc kubenswrapper[4953]: I1203 00:27:01.896457 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/5cd4ce9d-4089-4482-8910-83128307799d-buildworkdir\") pod \"5cd4ce9d-4089-4482-8910-83128307799d\" (UID: \"5cd4ce9d-4089-4482-8910-83128307799d\") " Dec 03 00:27:01 crc kubenswrapper[4953]: I1203 00:27:01.896486 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5cd4ce9d-4089-4482-8910-83128307799d-node-pullsecrets" (OuterVolumeSpecName: "node-pullsecrets") pod "5cd4ce9d-4089-4482-8910-83128307799d" (UID: "5cd4ce9d-4089-4482-8910-83128307799d"). InnerVolumeSpecName "node-pullsecrets". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 00:27:01 crc kubenswrapper[4953]: I1203 00:27:01.896511 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/5cd4ce9d-4089-4482-8910-83128307799d-buildcachedir\") pod \"5cd4ce9d-4089-4482-8910-83128307799d\" (UID: \"5cd4ce9d-4089-4482-8910-83128307799d\") " Dec 03 00:27:01 crc kubenswrapper[4953]: I1203 00:27:01.896558 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5cd4ce9d-4089-4482-8910-83128307799d-buildcachedir" (OuterVolumeSpecName: "buildcachedir") pod "5cd4ce9d-4089-4482-8910-83128307799d" (UID: "5cd4ce9d-4089-4482-8910-83128307799d"). InnerVolumeSpecName "buildcachedir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 00:27:01 crc kubenswrapper[4953]: I1203 00:27:01.896613 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/5cd4ce9d-4089-4482-8910-83128307799d-build-blob-cache\") pod \"5cd4ce9d-4089-4482-8910-83128307799d\" (UID: \"5cd4ce9d-4089-4482-8910-83128307799d\") " Dec 03 00:27:01 crc kubenswrapper[4953]: I1203 00:27:01.896822 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/5cd4ce9d-4089-4482-8910-83128307799d-build-ca-bundles\") pod \"5cd4ce9d-4089-4482-8910-83128307799d\" (UID: \"5cd4ce9d-4089-4482-8910-83128307799d\") " Dec 03 00:27:01 crc kubenswrapper[4953]: I1203 00:27:01.897405 4953 reconciler_common.go:293] "Volume detached for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/5cd4ce9d-4089-4482-8910-83128307799d-container-storage-run\") on node \"crc\" DevicePath \"\"" Dec 03 00:27:01 crc kubenswrapper[4953]: I1203 00:27:01.897618 4953 reconciler_common.go:293] "Volume detached for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/5cd4ce9d-4089-4482-8910-83128307799d-node-pullsecrets\") on node \"crc\" DevicePath \"\"" Dec 03 00:27:01 crc kubenswrapper[4953]: I1203 00:27:01.897659 4953 reconciler_common.go:293] "Volume detached for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/5cd4ce9d-4089-4482-8910-83128307799d-buildcachedir\") on node \"crc\" DevicePath \"\"" Dec 03 00:27:01 crc kubenswrapper[4953]: I1203 00:27:01.897399 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5cd4ce9d-4089-4482-8910-83128307799d-build-system-configs" (OuterVolumeSpecName: "build-system-configs") pod "5cd4ce9d-4089-4482-8910-83128307799d" (UID: "5cd4ce9d-4089-4482-8910-83128307799d"). InnerVolumeSpecName "build-system-configs". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:27:01 crc kubenswrapper[4953]: I1203 00:27:01.897457 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5cd4ce9d-4089-4482-8910-83128307799d-build-proxy-ca-bundles" (OuterVolumeSpecName: "build-proxy-ca-bundles") pod "5cd4ce9d-4089-4482-8910-83128307799d" (UID: "5cd4ce9d-4089-4482-8910-83128307799d"). InnerVolumeSpecName "build-proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:27:01 crc kubenswrapper[4953]: I1203 00:27:01.897813 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5cd4ce9d-4089-4482-8910-83128307799d-buildworkdir" (OuterVolumeSpecName: "buildworkdir") pod "5cd4ce9d-4089-4482-8910-83128307799d" (UID: "5cd4ce9d-4089-4482-8910-83128307799d"). InnerVolumeSpecName "buildworkdir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 00:27:01 crc kubenswrapper[4953]: I1203 00:27:01.897894 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5cd4ce9d-4089-4482-8910-83128307799d-build-ca-bundles" (OuterVolumeSpecName: "build-ca-bundles") pod "5cd4ce9d-4089-4482-8910-83128307799d" (UID: "5cd4ce9d-4089-4482-8910-83128307799d"). InnerVolumeSpecName "build-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:27:01 crc kubenswrapper[4953]: I1203 00:27:01.901632 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5cd4ce9d-4089-4482-8910-83128307799d-builder-dockercfg-9prwd-pull" (OuterVolumeSpecName: "builder-dockercfg-9prwd-pull") pod "5cd4ce9d-4089-4482-8910-83128307799d" (UID: "5cd4ce9d-4089-4482-8910-83128307799d"). InnerVolumeSpecName "builder-dockercfg-9prwd-pull". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:27:01 crc kubenswrapper[4953]: I1203 00:27:01.903430 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5cd4ce9d-4089-4482-8910-83128307799d-builder-dockercfg-9prwd-push" (OuterVolumeSpecName: "builder-dockercfg-9prwd-push") pod "5cd4ce9d-4089-4482-8910-83128307799d" (UID: "5cd4ce9d-4089-4482-8910-83128307799d"). InnerVolumeSpecName "builder-dockercfg-9prwd-push". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:27:01 crc kubenswrapper[4953]: I1203 00:27:01.908340 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5cd4ce9d-4089-4482-8910-83128307799d-kube-api-access-7grbf" (OuterVolumeSpecName: "kube-api-access-7grbf") pod "5cd4ce9d-4089-4482-8910-83128307799d" (UID: "5cd4ce9d-4089-4482-8910-83128307799d"). InnerVolumeSpecName "kube-api-access-7grbf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:27:01 crc kubenswrapper[4953]: I1203 00:27:01.998394 4953 reconciler_common.go:293] "Volume detached for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/5cd4ce9d-4089-4482-8910-83128307799d-build-ca-bundles\") on node \"crc\" DevicePath \"\"" Dec 03 00:27:01 crc kubenswrapper[4953]: I1203 00:27:01.998425 4953 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7grbf\" (UniqueName: \"kubernetes.io/projected/5cd4ce9d-4089-4482-8910-83128307799d-kube-api-access-7grbf\") on node \"crc\" DevicePath \"\"" Dec 03 00:27:01 crc kubenswrapper[4953]: I1203 00:27:01.998439 4953 reconciler_common.go:293] "Volume detached for volume \"builder-dockercfg-9prwd-pull\" (UniqueName: \"kubernetes.io/secret/5cd4ce9d-4089-4482-8910-83128307799d-builder-dockercfg-9prwd-pull\") on node \"crc\" DevicePath \"\"" Dec 03 00:27:01 crc kubenswrapper[4953]: I1203 00:27:01.998451 4953 reconciler_common.go:293] "Volume detached for volume \"builder-dockercfg-9prwd-push\" (UniqueName: \"kubernetes.io/secret/5cd4ce9d-4089-4482-8910-83128307799d-builder-dockercfg-9prwd-push\") on node \"crc\" DevicePath \"\"" Dec 03 00:27:01 crc kubenswrapper[4953]: I1203 00:27:01.998463 4953 reconciler_common.go:293] "Volume detached for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/5cd4ce9d-4089-4482-8910-83128307799d-build-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Dec 03 00:27:01 crc kubenswrapper[4953]: I1203 00:27:01.998474 4953 reconciler_common.go:293] "Volume detached for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/5cd4ce9d-4089-4482-8910-83128307799d-build-system-configs\") on node \"crc\" DevicePath \"\"" Dec 03 00:27:01 crc kubenswrapper[4953]: I1203 00:27:01.998485 4953 reconciler_common.go:293] "Volume detached for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/5cd4ce9d-4089-4482-8910-83128307799d-buildworkdir\") on node \"crc\" DevicePath \"\"" Dec 03 00:27:02 crc kubenswrapper[4953]: I1203 00:27:02.045114 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5cd4ce9d-4089-4482-8910-83128307799d-build-blob-cache" (OuterVolumeSpecName: "build-blob-cache") pod "5cd4ce9d-4089-4482-8910-83128307799d" (UID: "5cd4ce9d-4089-4482-8910-83128307799d"). InnerVolumeSpecName "build-blob-cache". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 00:27:02 crc kubenswrapper[4953]: I1203 00:27:02.074407 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5cd4ce9d-4089-4482-8910-83128307799d-container-storage-root" (OuterVolumeSpecName: "container-storage-root") pod "5cd4ce9d-4089-4482-8910-83128307799d" (UID: "5cd4ce9d-4089-4482-8910-83128307799d"). InnerVolumeSpecName "container-storage-root". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 00:27:02 crc kubenswrapper[4953]: I1203 00:27:02.097699 4953 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_sg-core-1-build_5cd4ce9d-4089-4482-8910-83128307799d/docker-build/0.log" Dec 03 00:27:02 crc kubenswrapper[4953]: I1203 00:27:02.098407 4953 generic.go:334] "Generic (PLEG): container finished" podID="5cd4ce9d-4089-4482-8910-83128307799d" containerID="2d1f0cede27b9aa51a9df95b41c323e6e9330bd930a5170a0a3d629e5d2c9e36" exitCode=1 Dec 03 00:27:02 crc kubenswrapper[4953]: I1203 00:27:02.098462 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/sg-core-1-build" event={"ID":"5cd4ce9d-4089-4482-8910-83128307799d","Type":"ContainerDied","Data":"2d1f0cede27b9aa51a9df95b41c323e6e9330bd930a5170a0a3d629e5d2c9e36"} Dec 03 00:27:02 crc kubenswrapper[4953]: I1203 00:27:02.098511 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/sg-core-1-build" event={"ID":"5cd4ce9d-4089-4482-8910-83128307799d","Type":"ContainerDied","Data":"7fcf976e4389cb7ec43e26d5fabe6f6c3da8c70dfd1db50fe11c00a5cf9d3736"} Dec 03 00:27:02 crc kubenswrapper[4953]: I1203 00:27:02.098519 4953 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/sg-core-1-build" Dec 03 00:27:02 crc kubenswrapper[4953]: I1203 00:27:02.098541 4953 scope.go:117] "RemoveContainer" containerID="2d1f0cede27b9aa51a9df95b41c323e6e9330bd930a5170a0a3d629e5d2c9e36" Dec 03 00:27:02 crc kubenswrapper[4953]: I1203 00:27:02.099345 4953 reconciler_common.go:293] "Volume detached for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/5cd4ce9d-4089-4482-8910-83128307799d-container-storage-root\") on node \"crc\" DevicePath \"\"" Dec 03 00:27:02 crc kubenswrapper[4953]: I1203 00:27:02.099382 4953 reconciler_common.go:293] "Volume detached for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/5cd4ce9d-4089-4482-8910-83128307799d-build-blob-cache\") on node \"crc\" DevicePath \"\"" Dec 03 00:27:02 crc kubenswrapper[4953]: I1203 00:27:02.146245 4953 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["service-telemetry/sg-core-1-build"] Dec 03 00:27:02 crc kubenswrapper[4953]: I1203 00:27:02.148563 4953 scope.go:117] "RemoveContainer" containerID="b11cc40cc1f04c662d2d8e4e6cb6ca1448e0bd6a7c96c80f6b0ca1e0b74f3149" Dec 03 00:27:02 crc kubenswrapper[4953]: I1203 00:27:02.151218 4953 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["service-telemetry/sg-core-1-build"] Dec 03 00:27:02 crc kubenswrapper[4953]: I1203 00:27:02.178820 4953 scope.go:117] "RemoveContainer" containerID="2d1f0cede27b9aa51a9df95b41c323e6e9330bd930a5170a0a3d629e5d2c9e36" Dec 03 00:27:02 crc kubenswrapper[4953]: E1203 00:27:02.179378 4953 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2d1f0cede27b9aa51a9df95b41c323e6e9330bd930a5170a0a3d629e5d2c9e36\": container with ID starting with 2d1f0cede27b9aa51a9df95b41c323e6e9330bd930a5170a0a3d629e5d2c9e36 not found: ID does not exist" containerID="2d1f0cede27b9aa51a9df95b41c323e6e9330bd930a5170a0a3d629e5d2c9e36" Dec 03 00:27:02 crc kubenswrapper[4953]: I1203 00:27:02.179413 4953 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2d1f0cede27b9aa51a9df95b41c323e6e9330bd930a5170a0a3d629e5d2c9e36"} err="failed to get container status \"2d1f0cede27b9aa51a9df95b41c323e6e9330bd930a5170a0a3d629e5d2c9e36\": rpc error: code = NotFound desc = could not find container \"2d1f0cede27b9aa51a9df95b41c323e6e9330bd930a5170a0a3d629e5d2c9e36\": container with ID starting with 2d1f0cede27b9aa51a9df95b41c323e6e9330bd930a5170a0a3d629e5d2c9e36 not found: ID does not exist" Dec 03 00:27:02 crc kubenswrapper[4953]: I1203 00:27:02.179433 4953 scope.go:117] "RemoveContainer" containerID="b11cc40cc1f04c662d2d8e4e6cb6ca1448e0bd6a7c96c80f6b0ca1e0b74f3149" Dec 03 00:27:02 crc kubenswrapper[4953]: E1203 00:27:02.179787 4953 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b11cc40cc1f04c662d2d8e4e6cb6ca1448e0bd6a7c96c80f6b0ca1e0b74f3149\": container with ID starting with b11cc40cc1f04c662d2d8e4e6cb6ca1448e0bd6a7c96c80f6b0ca1e0b74f3149 not found: ID does not exist" containerID="b11cc40cc1f04c662d2d8e4e6cb6ca1448e0bd6a7c96c80f6b0ca1e0b74f3149" Dec 03 00:27:02 crc kubenswrapper[4953]: I1203 00:27:02.179828 4953 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b11cc40cc1f04c662d2d8e4e6cb6ca1448e0bd6a7c96c80f6b0ca1e0b74f3149"} err="failed to get container status \"b11cc40cc1f04c662d2d8e4e6cb6ca1448e0bd6a7c96c80f6b0ca1e0b74f3149\": rpc error: code = NotFound desc = could not find container \"b11cc40cc1f04c662d2d8e4e6cb6ca1448e0bd6a7c96c80f6b0ca1e0b74f3149\": container with ID starting with b11cc40cc1f04c662d2d8e4e6cb6ca1448e0bd6a7c96c80f6b0ca1e0b74f3149 not found: ID does not exist" Dec 03 00:27:02 crc kubenswrapper[4953]: I1203 00:27:02.822812 4953 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["service-telemetry/sg-core-2-build"] Dec 03 00:27:02 crc kubenswrapper[4953]: E1203 00:27:02.823106 4953 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5cd4ce9d-4089-4482-8910-83128307799d" containerName="docker-build" Dec 03 00:27:02 crc kubenswrapper[4953]: I1203 00:27:02.823121 4953 state_mem.go:107] "Deleted CPUSet assignment" podUID="5cd4ce9d-4089-4482-8910-83128307799d" containerName="docker-build" Dec 03 00:27:02 crc kubenswrapper[4953]: E1203 00:27:02.823133 4953 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5cd4ce9d-4089-4482-8910-83128307799d" containerName="manage-dockerfile" Dec 03 00:27:02 crc kubenswrapper[4953]: I1203 00:27:02.823141 4953 state_mem.go:107] "Deleted CPUSet assignment" podUID="5cd4ce9d-4089-4482-8910-83128307799d" containerName="manage-dockerfile" Dec 03 00:27:02 crc kubenswrapper[4953]: I1203 00:27:02.823286 4953 memory_manager.go:354] "RemoveStaleState removing state" podUID="5cd4ce9d-4089-4482-8910-83128307799d" containerName="docker-build" Dec 03 00:27:02 crc kubenswrapper[4953]: I1203 00:27:02.824301 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/sg-core-2-build" Dec 03 00:27:02 crc kubenswrapper[4953]: I1203 00:27:02.828931 4953 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"builder-dockercfg-9prwd" Dec 03 00:27:02 crc kubenswrapper[4953]: I1203 00:27:02.829051 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"sg-core-2-global-ca" Dec 03 00:27:02 crc kubenswrapper[4953]: I1203 00:27:02.829102 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"sg-core-2-ca" Dec 03 00:27:02 crc kubenswrapper[4953]: I1203 00:27:02.829160 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"sg-core-2-sys-config" Dec 03 00:27:02 crc kubenswrapper[4953]: I1203 00:27:02.871028 4953 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/sg-core-2-build"] Dec 03 00:27:02 crc kubenswrapper[4953]: I1203 00:27:02.912323 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"builder-dockercfg-9prwd-push\" (UniqueName: \"kubernetes.io/secret/2a08a655-1dec-478f-ac65-ef93056e82bb-builder-dockercfg-9prwd-push\") pod \"sg-core-2-build\" (UID: \"2a08a655-1dec-478f-ac65-ef93056e82bb\") " pod="service-telemetry/sg-core-2-build" Dec 03 00:27:02 crc kubenswrapper[4953]: I1203 00:27:02.912368 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/2a08a655-1dec-478f-ac65-ef93056e82bb-container-storage-run\") pod \"sg-core-2-build\" (UID: \"2a08a655-1dec-478f-ac65-ef93056e82bb\") " pod="service-telemetry/sg-core-2-build" Dec 03 00:27:02 crc kubenswrapper[4953]: I1203 00:27:02.912408 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"builder-dockercfg-9prwd-pull\" (UniqueName: \"kubernetes.io/secret/2a08a655-1dec-478f-ac65-ef93056e82bb-builder-dockercfg-9prwd-pull\") pod \"sg-core-2-build\" (UID: \"2a08a655-1dec-478f-ac65-ef93056e82bb\") " pod="service-telemetry/sg-core-2-build" Dec 03 00:27:02 crc kubenswrapper[4953]: I1203 00:27:02.912431 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/2a08a655-1dec-478f-ac65-ef93056e82bb-buildcachedir\") pod \"sg-core-2-build\" (UID: \"2a08a655-1dec-478f-ac65-ef93056e82bb\") " pod="service-telemetry/sg-core-2-build" Dec 03 00:27:02 crc kubenswrapper[4953]: I1203 00:27:02.912458 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/2a08a655-1dec-478f-ac65-ef93056e82bb-node-pullsecrets\") pod \"sg-core-2-build\" (UID: \"2a08a655-1dec-478f-ac65-ef93056e82bb\") " pod="service-telemetry/sg-core-2-build" Dec 03 00:27:02 crc kubenswrapper[4953]: I1203 00:27:02.912482 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/2a08a655-1dec-478f-ac65-ef93056e82bb-build-blob-cache\") pod \"sg-core-2-build\" (UID: \"2a08a655-1dec-478f-ac65-ef93056e82bb\") " pod="service-telemetry/sg-core-2-build" Dec 03 00:27:02 crc kubenswrapper[4953]: I1203 00:27:02.912576 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/2a08a655-1dec-478f-ac65-ef93056e82bb-buildworkdir\") pod \"sg-core-2-build\" (UID: \"2a08a655-1dec-478f-ac65-ef93056e82bb\") " pod="service-telemetry/sg-core-2-build" Dec 03 00:27:02 crc kubenswrapper[4953]: I1203 00:27:02.912639 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/2a08a655-1dec-478f-ac65-ef93056e82bb-build-ca-bundles\") pod \"sg-core-2-build\" (UID: \"2a08a655-1dec-478f-ac65-ef93056e82bb\") " pod="service-telemetry/sg-core-2-build" Dec 03 00:27:02 crc kubenswrapper[4953]: I1203 00:27:02.912668 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/2a08a655-1dec-478f-ac65-ef93056e82bb-build-proxy-ca-bundles\") pod \"sg-core-2-build\" (UID: \"2a08a655-1dec-478f-ac65-ef93056e82bb\") " pod="service-telemetry/sg-core-2-build" Dec 03 00:27:02 crc kubenswrapper[4953]: I1203 00:27:02.912775 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2bf8k\" (UniqueName: \"kubernetes.io/projected/2a08a655-1dec-478f-ac65-ef93056e82bb-kube-api-access-2bf8k\") pod \"sg-core-2-build\" (UID: \"2a08a655-1dec-478f-ac65-ef93056e82bb\") " pod="service-telemetry/sg-core-2-build" Dec 03 00:27:02 crc kubenswrapper[4953]: I1203 00:27:02.912816 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/2a08a655-1dec-478f-ac65-ef93056e82bb-container-storage-root\") pod \"sg-core-2-build\" (UID: \"2a08a655-1dec-478f-ac65-ef93056e82bb\") " pod="service-telemetry/sg-core-2-build" Dec 03 00:27:02 crc kubenswrapper[4953]: I1203 00:27:02.912862 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/2a08a655-1dec-478f-ac65-ef93056e82bb-build-system-configs\") pod \"sg-core-2-build\" (UID: \"2a08a655-1dec-478f-ac65-ef93056e82bb\") " pod="service-telemetry/sg-core-2-build" Dec 03 00:27:03 crc kubenswrapper[4953]: I1203 00:27:03.013917 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"builder-dockercfg-9prwd-push\" (UniqueName: \"kubernetes.io/secret/2a08a655-1dec-478f-ac65-ef93056e82bb-builder-dockercfg-9prwd-push\") pod \"sg-core-2-build\" (UID: \"2a08a655-1dec-478f-ac65-ef93056e82bb\") " pod="service-telemetry/sg-core-2-build" Dec 03 00:27:03 crc kubenswrapper[4953]: I1203 00:27:03.013964 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/2a08a655-1dec-478f-ac65-ef93056e82bb-container-storage-run\") pod \"sg-core-2-build\" (UID: \"2a08a655-1dec-478f-ac65-ef93056e82bb\") " pod="service-telemetry/sg-core-2-build" Dec 03 00:27:03 crc kubenswrapper[4953]: I1203 00:27:03.014028 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"builder-dockercfg-9prwd-pull\" (UniqueName: \"kubernetes.io/secret/2a08a655-1dec-478f-ac65-ef93056e82bb-builder-dockercfg-9prwd-pull\") pod \"sg-core-2-build\" (UID: \"2a08a655-1dec-478f-ac65-ef93056e82bb\") " pod="service-telemetry/sg-core-2-build" Dec 03 00:27:03 crc kubenswrapper[4953]: I1203 00:27:03.014053 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/2a08a655-1dec-478f-ac65-ef93056e82bb-buildcachedir\") pod \"sg-core-2-build\" (UID: \"2a08a655-1dec-478f-ac65-ef93056e82bb\") " pod="service-telemetry/sg-core-2-build" Dec 03 00:27:03 crc kubenswrapper[4953]: I1203 00:27:03.014085 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/2a08a655-1dec-478f-ac65-ef93056e82bb-node-pullsecrets\") pod \"sg-core-2-build\" (UID: \"2a08a655-1dec-478f-ac65-ef93056e82bb\") " pod="service-telemetry/sg-core-2-build" Dec 03 00:27:03 crc kubenswrapper[4953]: I1203 00:27:03.014111 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/2a08a655-1dec-478f-ac65-ef93056e82bb-build-blob-cache\") pod \"sg-core-2-build\" (UID: \"2a08a655-1dec-478f-ac65-ef93056e82bb\") " pod="service-telemetry/sg-core-2-build" Dec 03 00:27:03 crc kubenswrapper[4953]: I1203 00:27:03.014137 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/2a08a655-1dec-478f-ac65-ef93056e82bb-buildworkdir\") pod \"sg-core-2-build\" (UID: \"2a08a655-1dec-478f-ac65-ef93056e82bb\") " pod="service-telemetry/sg-core-2-build" Dec 03 00:27:03 crc kubenswrapper[4953]: I1203 00:27:03.014159 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/2a08a655-1dec-478f-ac65-ef93056e82bb-build-ca-bundles\") pod \"sg-core-2-build\" (UID: \"2a08a655-1dec-478f-ac65-ef93056e82bb\") " pod="service-telemetry/sg-core-2-build" Dec 03 00:27:03 crc kubenswrapper[4953]: I1203 00:27:03.014181 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/2a08a655-1dec-478f-ac65-ef93056e82bb-build-proxy-ca-bundles\") pod \"sg-core-2-build\" (UID: \"2a08a655-1dec-478f-ac65-ef93056e82bb\") " pod="service-telemetry/sg-core-2-build" Dec 03 00:27:03 crc kubenswrapper[4953]: I1203 00:27:03.014221 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2bf8k\" (UniqueName: \"kubernetes.io/projected/2a08a655-1dec-478f-ac65-ef93056e82bb-kube-api-access-2bf8k\") pod \"sg-core-2-build\" (UID: \"2a08a655-1dec-478f-ac65-ef93056e82bb\") " pod="service-telemetry/sg-core-2-build" Dec 03 00:27:03 crc kubenswrapper[4953]: I1203 00:27:03.014248 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/2a08a655-1dec-478f-ac65-ef93056e82bb-container-storage-root\") pod \"sg-core-2-build\" (UID: \"2a08a655-1dec-478f-ac65-ef93056e82bb\") " pod="service-telemetry/sg-core-2-build" Dec 03 00:27:03 crc kubenswrapper[4953]: I1203 00:27:03.014277 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/2a08a655-1dec-478f-ac65-ef93056e82bb-build-system-configs\") pod \"sg-core-2-build\" (UID: \"2a08a655-1dec-478f-ac65-ef93056e82bb\") " pod="service-telemetry/sg-core-2-build" Dec 03 00:27:03 crc kubenswrapper[4953]: I1203 00:27:03.014536 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/2a08a655-1dec-478f-ac65-ef93056e82bb-container-storage-run\") pod \"sg-core-2-build\" (UID: \"2a08a655-1dec-478f-ac65-ef93056e82bb\") " pod="service-telemetry/sg-core-2-build" Dec 03 00:27:03 crc kubenswrapper[4953]: I1203 00:27:03.014824 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/2a08a655-1dec-478f-ac65-ef93056e82bb-buildworkdir\") pod \"sg-core-2-build\" (UID: \"2a08a655-1dec-478f-ac65-ef93056e82bb\") " pod="service-telemetry/sg-core-2-build" Dec 03 00:27:03 crc kubenswrapper[4953]: I1203 00:27:03.014919 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/2a08a655-1dec-478f-ac65-ef93056e82bb-build-system-configs\") pod \"sg-core-2-build\" (UID: \"2a08a655-1dec-478f-ac65-ef93056e82bb\") " pod="service-telemetry/sg-core-2-build" Dec 03 00:27:03 crc kubenswrapper[4953]: I1203 00:27:03.014996 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/2a08a655-1dec-478f-ac65-ef93056e82bb-node-pullsecrets\") pod \"sg-core-2-build\" (UID: \"2a08a655-1dec-478f-ac65-ef93056e82bb\") " pod="service-telemetry/sg-core-2-build" Dec 03 00:27:03 crc kubenswrapper[4953]: I1203 00:27:03.015077 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/2a08a655-1dec-478f-ac65-ef93056e82bb-buildcachedir\") pod \"sg-core-2-build\" (UID: \"2a08a655-1dec-478f-ac65-ef93056e82bb\") " pod="service-telemetry/sg-core-2-build" Dec 03 00:27:03 crc kubenswrapper[4953]: I1203 00:27:03.015203 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/2a08a655-1dec-478f-ac65-ef93056e82bb-container-storage-root\") pod \"sg-core-2-build\" (UID: \"2a08a655-1dec-478f-ac65-ef93056e82bb\") " pod="service-telemetry/sg-core-2-build" Dec 03 00:27:03 crc kubenswrapper[4953]: I1203 00:27:03.015214 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/2a08a655-1dec-478f-ac65-ef93056e82bb-build-proxy-ca-bundles\") pod \"sg-core-2-build\" (UID: \"2a08a655-1dec-478f-ac65-ef93056e82bb\") " pod="service-telemetry/sg-core-2-build" Dec 03 00:27:03 crc kubenswrapper[4953]: I1203 00:27:03.015432 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/2a08a655-1dec-478f-ac65-ef93056e82bb-build-blob-cache\") pod \"sg-core-2-build\" (UID: \"2a08a655-1dec-478f-ac65-ef93056e82bb\") " pod="service-telemetry/sg-core-2-build" Dec 03 00:27:03 crc kubenswrapper[4953]: I1203 00:27:03.016426 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/2a08a655-1dec-478f-ac65-ef93056e82bb-build-ca-bundles\") pod \"sg-core-2-build\" (UID: \"2a08a655-1dec-478f-ac65-ef93056e82bb\") " pod="service-telemetry/sg-core-2-build" Dec 03 00:27:03 crc kubenswrapper[4953]: I1203 00:27:03.018183 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"builder-dockercfg-9prwd-push\" (UniqueName: \"kubernetes.io/secret/2a08a655-1dec-478f-ac65-ef93056e82bb-builder-dockercfg-9prwd-push\") pod \"sg-core-2-build\" (UID: \"2a08a655-1dec-478f-ac65-ef93056e82bb\") " pod="service-telemetry/sg-core-2-build" Dec 03 00:27:03 crc kubenswrapper[4953]: I1203 00:27:03.018183 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"builder-dockercfg-9prwd-pull\" (UniqueName: \"kubernetes.io/secret/2a08a655-1dec-478f-ac65-ef93056e82bb-builder-dockercfg-9prwd-pull\") pod \"sg-core-2-build\" (UID: \"2a08a655-1dec-478f-ac65-ef93056e82bb\") " pod="service-telemetry/sg-core-2-build" Dec 03 00:27:03 crc kubenswrapper[4953]: I1203 00:27:03.030572 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2bf8k\" (UniqueName: \"kubernetes.io/projected/2a08a655-1dec-478f-ac65-ef93056e82bb-kube-api-access-2bf8k\") pod \"sg-core-2-build\" (UID: \"2a08a655-1dec-478f-ac65-ef93056e82bb\") " pod="service-telemetry/sg-core-2-build" Dec 03 00:27:03 crc kubenswrapper[4953]: I1203 00:27:03.175295 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/sg-core-2-build" Dec 03 00:27:03 crc kubenswrapper[4953]: I1203 00:27:03.357874 4953 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5cd4ce9d-4089-4482-8910-83128307799d" path="/var/lib/kubelet/pods/5cd4ce9d-4089-4482-8910-83128307799d/volumes" Dec 03 00:27:03 crc kubenswrapper[4953]: I1203 00:27:03.408872 4953 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/sg-core-2-build"] Dec 03 00:27:04 crc kubenswrapper[4953]: I1203 00:27:04.117594 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/sg-core-2-build" event={"ID":"2a08a655-1dec-478f-ac65-ef93056e82bb","Type":"ContainerStarted","Data":"2ebd93e6a7af5cfe38430109bf0be30647a51e1694650f27e71fb2bbec2fd1af"} Dec 03 00:27:04 crc kubenswrapper[4953]: I1203 00:27:04.117920 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/sg-core-2-build" event={"ID":"2a08a655-1dec-478f-ac65-ef93056e82bb","Type":"ContainerStarted","Data":"23d5a199350d200228f309ef2bcc1454f9d8956d2234a2d6ca304b76cc029e38"} Dec 03 00:27:05 crc kubenswrapper[4953]: I1203 00:27:05.132172 4953 generic.go:334] "Generic (PLEG): container finished" podID="2a08a655-1dec-478f-ac65-ef93056e82bb" containerID="2ebd93e6a7af5cfe38430109bf0be30647a51e1694650f27e71fb2bbec2fd1af" exitCode=0 Dec 03 00:27:05 crc kubenswrapper[4953]: I1203 00:27:05.132222 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/sg-core-2-build" event={"ID":"2a08a655-1dec-478f-ac65-ef93056e82bb","Type":"ContainerDied","Data":"2ebd93e6a7af5cfe38430109bf0be30647a51e1694650f27e71fb2bbec2fd1af"} Dec 03 00:27:06 crc kubenswrapper[4953]: I1203 00:27:06.141754 4953 generic.go:334] "Generic (PLEG): container finished" podID="2a08a655-1dec-478f-ac65-ef93056e82bb" containerID="af39150055081f6f7f15894997aecc8137a245161a283d738d5114ae6ab4c5ed" exitCode=0 Dec 03 00:27:06 crc kubenswrapper[4953]: I1203 00:27:06.141830 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/sg-core-2-build" event={"ID":"2a08a655-1dec-478f-ac65-ef93056e82bb","Type":"ContainerDied","Data":"af39150055081f6f7f15894997aecc8137a245161a283d738d5114ae6ab4c5ed"} Dec 03 00:27:06 crc kubenswrapper[4953]: I1203 00:27:06.200025 4953 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_sg-core-2-build_2a08a655-1dec-478f-ac65-ef93056e82bb/manage-dockerfile/0.log" Dec 03 00:27:07 crc kubenswrapper[4953]: I1203 00:27:07.153599 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/sg-core-2-build" event={"ID":"2a08a655-1dec-478f-ac65-ef93056e82bb","Type":"ContainerStarted","Data":"923108362fce543c78af613faa4bf277b3f846cd54fd21513215687364bef773"} Dec 03 00:27:07 crc kubenswrapper[4953]: I1203 00:27:07.201333 4953 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="service-telemetry/sg-core-2-build" podStartSLOduration=5.201307165 podStartE2EDuration="5.201307165s" podCreationTimestamp="2025-12-03 00:27:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 00:27:07.193121511 +0000 UTC m=+1216.177402324" watchObservedRunningTime="2025-12-03 00:27:07.201307165 +0000 UTC m=+1216.185587998" Dec 03 00:27:18 crc kubenswrapper[4953]: I1203 00:27:18.944249 4953 patch_prober.go:28] interesting pod/machine-config-daemon-6dfbn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 00:27:18 crc kubenswrapper[4953]: I1203 00:27:18.944692 4953 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-6dfbn" podUID="53ffb1e9-9a2e-4869-80d7-776f4aa7ef8b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 00:27:18 crc kubenswrapper[4953]: I1203 00:27:18.944740 4953 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-6dfbn" Dec 03 00:27:18 crc kubenswrapper[4953]: I1203 00:27:18.945317 4953 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"1ed646ac60a808beb2f58a1886a69c6cb3f62ef93231fc4501030f895ee3958c"} pod="openshift-machine-config-operator/machine-config-daemon-6dfbn" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 03 00:27:18 crc kubenswrapper[4953]: I1203 00:27:18.945369 4953 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-6dfbn" podUID="53ffb1e9-9a2e-4869-80d7-776f4aa7ef8b" containerName="machine-config-daemon" containerID="cri-o://1ed646ac60a808beb2f58a1886a69c6cb3f62ef93231fc4501030f895ee3958c" gracePeriod=600 Dec 03 00:27:19 crc kubenswrapper[4953]: I1203 00:27:19.237418 4953 generic.go:334] "Generic (PLEG): container finished" podID="53ffb1e9-9a2e-4869-80d7-776f4aa7ef8b" containerID="1ed646ac60a808beb2f58a1886a69c6cb3f62ef93231fc4501030f895ee3958c" exitCode=0 Dec 03 00:27:19 crc kubenswrapper[4953]: I1203 00:27:19.237466 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-6dfbn" event={"ID":"53ffb1e9-9a2e-4869-80d7-776f4aa7ef8b","Type":"ContainerDied","Data":"1ed646ac60a808beb2f58a1886a69c6cb3f62ef93231fc4501030f895ee3958c"} Dec 03 00:27:19 crc kubenswrapper[4953]: I1203 00:27:19.237512 4953 scope.go:117] "RemoveContainer" containerID="ede43397367a8a468be56d98b9f73d5afe6a5294be1471b33f69b64147ab3995" Dec 03 00:27:20 crc kubenswrapper[4953]: I1203 00:27:20.246529 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-6dfbn" event={"ID":"53ffb1e9-9a2e-4869-80d7-776f4aa7ef8b","Type":"ContainerStarted","Data":"995d5837fd18ae79a3033d54ed80db8a681ac98304a63c306fc50ba95ea36238"} Dec 03 00:29:48 crc kubenswrapper[4953]: I1203 00:29:48.944154 4953 patch_prober.go:28] interesting pod/machine-config-daemon-6dfbn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 00:29:48 crc kubenswrapper[4953]: I1203 00:29:48.944753 4953 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-6dfbn" podUID="53ffb1e9-9a2e-4869-80d7-776f4aa7ef8b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 00:30:00 crc kubenswrapper[4953]: I1203 00:30:00.149583 4953 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412030-k9kxc"] Dec 03 00:30:00 crc kubenswrapper[4953]: I1203 00:30:00.151237 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29412030-k9kxc" Dec 03 00:30:00 crc kubenswrapper[4953]: I1203 00:30:00.154350 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 03 00:30:00 crc kubenswrapper[4953]: I1203 00:30:00.155469 4953 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 03 00:30:00 crc kubenswrapper[4953]: I1203 00:30:00.163946 4953 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412030-k9kxc"] Dec 03 00:30:00 crc kubenswrapper[4953]: I1203 00:30:00.246330 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/dad9c3d8-bc04-466b-9d82-4120eb7e8139-config-volume\") pod \"collect-profiles-29412030-k9kxc\" (UID: \"dad9c3d8-bc04-466b-9d82-4120eb7e8139\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412030-k9kxc" Dec 03 00:30:00 crc kubenswrapper[4953]: I1203 00:30:00.246387 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/dad9c3d8-bc04-466b-9d82-4120eb7e8139-secret-volume\") pod \"collect-profiles-29412030-k9kxc\" (UID: \"dad9c3d8-bc04-466b-9d82-4120eb7e8139\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412030-k9kxc" Dec 03 00:30:00 crc kubenswrapper[4953]: I1203 00:30:00.246429 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xz7b6\" (UniqueName: \"kubernetes.io/projected/dad9c3d8-bc04-466b-9d82-4120eb7e8139-kube-api-access-xz7b6\") pod \"collect-profiles-29412030-k9kxc\" (UID: \"dad9c3d8-bc04-466b-9d82-4120eb7e8139\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412030-k9kxc" Dec 03 00:30:00 crc kubenswrapper[4953]: I1203 00:30:00.347872 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/dad9c3d8-bc04-466b-9d82-4120eb7e8139-config-volume\") pod \"collect-profiles-29412030-k9kxc\" (UID: \"dad9c3d8-bc04-466b-9d82-4120eb7e8139\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412030-k9kxc" Dec 03 00:30:00 crc kubenswrapper[4953]: I1203 00:30:00.347924 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/dad9c3d8-bc04-466b-9d82-4120eb7e8139-secret-volume\") pod \"collect-profiles-29412030-k9kxc\" (UID: \"dad9c3d8-bc04-466b-9d82-4120eb7e8139\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412030-k9kxc" Dec 03 00:30:00 crc kubenswrapper[4953]: I1203 00:30:00.347966 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xz7b6\" (UniqueName: \"kubernetes.io/projected/dad9c3d8-bc04-466b-9d82-4120eb7e8139-kube-api-access-xz7b6\") pod \"collect-profiles-29412030-k9kxc\" (UID: \"dad9c3d8-bc04-466b-9d82-4120eb7e8139\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412030-k9kxc" Dec 03 00:30:00 crc kubenswrapper[4953]: I1203 00:30:00.357378 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/dad9c3d8-bc04-466b-9d82-4120eb7e8139-config-volume\") pod \"collect-profiles-29412030-k9kxc\" (UID: \"dad9c3d8-bc04-466b-9d82-4120eb7e8139\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412030-k9kxc" Dec 03 00:30:00 crc kubenswrapper[4953]: I1203 00:30:00.375351 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xz7b6\" (UniqueName: \"kubernetes.io/projected/dad9c3d8-bc04-466b-9d82-4120eb7e8139-kube-api-access-xz7b6\") pod \"collect-profiles-29412030-k9kxc\" (UID: \"dad9c3d8-bc04-466b-9d82-4120eb7e8139\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412030-k9kxc" Dec 03 00:30:00 crc kubenswrapper[4953]: I1203 00:30:00.380028 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/dad9c3d8-bc04-466b-9d82-4120eb7e8139-secret-volume\") pod \"collect-profiles-29412030-k9kxc\" (UID: \"dad9c3d8-bc04-466b-9d82-4120eb7e8139\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412030-k9kxc" Dec 03 00:30:00 crc kubenswrapper[4953]: I1203 00:30:00.470680 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29412030-k9kxc" Dec 03 00:30:00 crc kubenswrapper[4953]: I1203 00:30:00.726678 4953 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412030-k9kxc"] Dec 03 00:30:00 crc kubenswrapper[4953]: I1203 00:30:00.949172 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29412030-k9kxc" event={"ID":"dad9c3d8-bc04-466b-9d82-4120eb7e8139","Type":"ContainerStarted","Data":"d28ae2d19bf8ceb2dd20b65ddcf3c98be2363ba1b9a3ffb0140a80633473acee"} Dec 03 00:30:01 crc kubenswrapper[4953]: I1203 00:30:01.958498 4953 generic.go:334] "Generic (PLEG): container finished" podID="dad9c3d8-bc04-466b-9d82-4120eb7e8139" containerID="473ebe7a7a2d2e8b50f639b49859542213fa2fa57b9222f54ed39304ff928518" exitCode=0 Dec 03 00:30:01 crc kubenswrapper[4953]: I1203 00:30:01.958567 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29412030-k9kxc" event={"ID":"dad9c3d8-bc04-466b-9d82-4120eb7e8139","Type":"ContainerDied","Data":"473ebe7a7a2d2e8b50f639b49859542213fa2fa57b9222f54ed39304ff928518"} Dec 03 00:30:03 crc kubenswrapper[4953]: I1203 00:30:03.453245 4953 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29412030-k9kxc" Dec 03 00:30:03 crc kubenswrapper[4953]: I1203 00:30:03.619392 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/dad9c3d8-bc04-466b-9d82-4120eb7e8139-secret-volume\") pod \"dad9c3d8-bc04-466b-9d82-4120eb7e8139\" (UID: \"dad9c3d8-bc04-466b-9d82-4120eb7e8139\") " Dec 03 00:30:03 crc kubenswrapper[4953]: I1203 00:30:03.619603 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/dad9c3d8-bc04-466b-9d82-4120eb7e8139-config-volume\") pod \"dad9c3d8-bc04-466b-9d82-4120eb7e8139\" (UID: \"dad9c3d8-bc04-466b-9d82-4120eb7e8139\") " Dec 03 00:30:03 crc kubenswrapper[4953]: I1203 00:30:03.619684 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xz7b6\" (UniqueName: \"kubernetes.io/projected/dad9c3d8-bc04-466b-9d82-4120eb7e8139-kube-api-access-xz7b6\") pod \"dad9c3d8-bc04-466b-9d82-4120eb7e8139\" (UID: \"dad9c3d8-bc04-466b-9d82-4120eb7e8139\") " Dec 03 00:30:03 crc kubenswrapper[4953]: I1203 00:30:03.620935 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dad9c3d8-bc04-466b-9d82-4120eb7e8139-config-volume" (OuterVolumeSpecName: "config-volume") pod "dad9c3d8-bc04-466b-9d82-4120eb7e8139" (UID: "dad9c3d8-bc04-466b-9d82-4120eb7e8139"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:30:03 crc kubenswrapper[4953]: I1203 00:30:03.763885 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dad9c3d8-bc04-466b-9d82-4120eb7e8139-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "dad9c3d8-bc04-466b-9d82-4120eb7e8139" (UID: "dad9c3d8-bc04-466b-9d82-4120eb7e8139"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:30:03 crc kubenswrapper[4953]: I1203 00:30:03.764414 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dad9c3d8-bc04-466b-9d82-4120eb7e8139-kube-api-access-xz7b6" (OuterVolumeSpecName: "kube-api-access-xz7b6") pod "dad9c3d8-bc04-466b-9d82-4120eb7e8139" (UID: "dad9c3d8-bc04-466b-9d82-4120eb7e8139"). InnerVolumeSpecName "kube-api-access-xz7b6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:30:03 crc kubenswrapper[4953]: I1203 00:30:03.764426 4953 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/dad9c3d8-bc04-466b-9d82-4120eb7e8139-config-volume\") on node \"crc\" DevicePath \"\"" Dec 03 00:30:03 crc kubenswrapper[4953]: I1203 00:30:03.764468 4953 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/dad9c3d8-bc04-466b-9d82-4120eb7e8139-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 03 00:30:03 crc kubenswrapper[4953]: I1203 00:30:03.865286 4953 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xz7b6\" (UniqueName: \"kubernetes.io/projected/dad9c3d8-bc04-466b-9d82-4120eb7e8139-kube-api-access-xz7b6\") on node \"crc\" DevicePath \"\"" Dec 03 00:30:04 crc kubenswrapper[4953]: I1203 00:30:03.996675 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29412030-k9kxc" event={"ID":"dad9c3d8-bc04-466b-9d82-4120eb7e8139","Type":"ContainerDied","Data":"d28ae2d19bf8ceb2dd20b65ddcf3c98be2363ba1b9a3ffb0140a80633473acee"} Dec 03 00:30:04 crc kubenswrapper[4953]: I1203 00:30:03.996783 4953 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d28ae2d19bf8ceb2dd20b65ddcf3c98be2363ba1b9a3ffb0140a80633473acee" Dec 03 00:30:04 crc kubenswrapper[4953]: I1203 00:30:03.996933 4953 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29412030-k9kxc" Dec 03 00:30:18 crc kubenswrapper[4953]: I1203 00:30:18.944438 4953 patch_prober.go:28] interesting pod/machine-config-daemon-6dfbn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 00:30:18 crc kubenswrapper[4953]: I1203 00:30:18.945321 4953 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-6dfbn" podUID="53ffb1e9-9a2e-4869-80d7-776f4aa7ef8b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 00:30:48 crc kubenswrapper[4953]: I1203 00:30:48.944556 4953 patch_prober.go:28] interesting pod/machine-config-daemon-6dfbn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 00:30:48 crc kubenswrapper[4953]: I1203 00:30:48.945580 4953 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-6dfbn" podUID="53ffb1e9-9a2e-4869-80d7-776f4aa7ef8b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 00:30:48 crc kubenswrapper[4953]: I1203 00:30:48.945679 4953 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-6dfbn" Dec 03 00:30:48 crc kubenswrapper[4953]: I1203 00:30:48.946741 4953 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"995d5837fd18ae79a3033d54ed80db8a681ac98304a63c306fc50ba95ea36238"} pod="openshift-machine-config-operator/machine-config-daemon-6dfbn" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 03 00:30:48 crc kubenswrapper[4953]: I1203 00:30:48.946818 4953 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-6dfbn" podUID="53ffb1e9-9a2e-4869-80d7-776f4aa7ef8b" containerName="machine-config-daemon" containerID="cri-o://995d5837fd18ae79a3033d54ed80db8a681ac98304a63c306fc50ba95ea36238" gracePeriod=600 Dec 03 00:30:49 crc kubenswrapper[4953]: I1203 00:30:49.759737 4953 generic.go:334] "Generic (PLEG): container finished" podID="53ffb1e9-9a2e-4869-80d7-776f4aa7ef8b" containerID="995d5837fd18ae79a3033d54ed80db8a681ac98304a63c306fc50ba95ea36238" exitCode=0 Dec 03 00:30:49 crc kubenswrapper[4953]: I1203 00:30:49.759842 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-6dfbn" event={"ID":"53ffb1e9-9a2e-4869-80d7-776f4aa7ef8b","Type":"ContainerDied","Data":"995d5837fd18ae79a3033d54ed80db8a681ac98304a63c306fc50ba95ea36238"} Dec 03 00:30:49 crc kubenswrapper[4953]: I1203 00:30:49.760178 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-6dfbn" event={"ID":"53ffb1e9-9a2e-4869-80d7-776f4aa7ef8b","Type":"ContainerStarted","Data":"150ffc325f215c6da77fc72304c13cb10dafc800bbc932afd319cc34442a3f76"} Dec 03 00:30:49 crc kubenswrapper[4953]: I1203 00:30:49.760206 4953 scope.go:117] "RemoveContainer" containerID="1ed646ac60a808beb2f58a1886a69c6cb3f62ef93231fc4501030f895ee3958c" Dec 03 00:31:01 crc kubenswrapper[4953]: I1203 00:31:01.850802 4953 generic.go:334] "Generic (PLEG): container finished" podID="2a08a655-1dec-478f-ac65-ef93056e82bb" containerID="923108362fce543c78af613faa4bf277b3f846cd54fd21513215687364bef773" exitCode=0 Dec 03 00:31:01 crc kubenswrapper[4953]: I1203 00:31:01.850894 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/sg-core-2-build" event={"ID":"2a08a655-1dec-478f-ac65-ef93056e82bb","Type":"ContainerDied","Data":"923108362fce543c78af613faa4bf277b3f846cd54fd21513215687364bef773"} Dec 03 00:31:03 crc kubenswrapper[4953]: I1203 00:31:03.101875 4953 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/sg-core-2-build" Dec 03 00:31:03 crc kubenswrapper[4953]: I1203 00:31:03.236037 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/2a08a655-1dec-478f-ac65-ef93056e82bb-build-ca-bundles\") pod \"2a08a655-1dec-478f-ac65-ef93056e82bb\" (UID: \"2a08a655-1dec-478f-ac65-ef93056e82bb\") " Dec 03 00:31:03 crc kubenswrapper[4953]: I1203 00:31:03.236177 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/2a08a655-1dec-478f-ac65-ef93056e82bb-buildcachedir\") pod \"2a08a655-1dec-478f-ac65-ef93056e82bb\" (UID: \"2a08a655-1dec-478f-ac65-ef93056e82bb\") " Dec 03 00:31:03 crc kubenswrapper[4953]: I1203 00:31:03.236216 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/2a08a655-1dec-478f-ac65-ef93056e82bb-buildworkdir\") pod \"2a08a655-1dec-478f-ac65-ef93056e82bb\" (UID: \"2a08a655-1dec-478f-ac65-ef93056e82bb\") " Dec 03 00:31:03 crc kubenswrapper[4953]: I1203 00:31:03.236323 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/2a08a655-1dec-478f-ac65-ef93056e82bb-build-blob-cache\") pod \"2a08a655-1dec-478f-ac65-ef93056e82bb\" (UID: \"2a08a655-1dec-478f-ac65-ef93056e82bb\") " Dec 03 00:31:03 crc kubenswrapper[4953]: I1203 00:31:03.236313 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2a08a655-1dec-478f-ac65-ef93056e82bb-buildcachedir" (OuterVolumeSpecName: "buildcachedir") pod "2a08a655-1dec-478f-ac65-ef93056e82bb" (UID: "2a08a655-1dec-478f-ac65-ef93056e82bb"). InnerVolumeSpecName "buildcachedir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 00:31:03 crc kubenswrapper[4953]: I1203 00:31:03.236360 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/2a08a655-1dec-478f-ac65-ef93056e82bb-container-storage-run\") pod \"2a08a655-1dec-478f-ac65-ef93056e82bb\" (UID: \"2a08a655-1dec-478f-ac65-ef93056e82bb\") " Dec 03 00:31:03 crc kubenswrapper[4953]: I1203 00:31:03.236395 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/2a08a655-1dec-478f-ac65-ef93056e82bb-container-storage-root\") pod \"2a08a655-1dec-478f-ac65-ef93056e82bb\" (UID: \"2a08a655-1dec-478f-ac65-ef93056e82bb\") " Dec 03 00:31:03 crc kubenswrapper[4953]: I1203 00:31:03.236414 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/2a08a655-1dec-478f-ac65-ef93056e82bb-node-pullsecrets\") pod \"2a08a655-1dec-478f-ac65-ef93056e82bb\" (UID: \"2a08a655-1dec-478f-ac65-ef93056e82bb\") " Dec 03 00:31:03 crc kubenswrapper[4953]: I1203 00:31:03.236441 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2bf8k\" (UniqueName: \"kubernetes.io/projected/2a08a655-1dec-478f-ac65-ef93056e82bb-kube-api-access-2bf8k\") pod \"2a08a655-1dec-478f-ac65-ef93056e82bb\" (UID: \"2a08a655-1dec-478f-ac65-ef93056e82bb\") " Dec 03 00:31:03 crc kubenswrapper[4953]: I1203 00:31:03.236521 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/2a08a655-1dec-478f-ac65-ef93056e82bb-build-system-configs\") pod \"2a08a655-1dec-478f-ac65-ef93056e82bb\" (UID: \"2a08a655-1dec-478f-ac65-ef93056e82bb\") " Dec 03 00:31:03 crc kubenswrapper[4953]: I1203 00:31:03.236542 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2a08a655-1dec-478f-ac65-ef93056e82bb-node-pullsecrets" (OuterVolumeSpecName: "node-pullsecrets") pod "2a08a655-1dec-478f-ac65-ef93056e82bb" (UID: "2a08a655-1dec-478f-ac65-ef93056e82bb"). InnerVolumeSpecName "node-pullsecrets". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 00:31:03 crc kubenswrapper[4953]: I1203 00:31:03.237686 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2a08a655-1dec-478f-ac65-ef93056e82bb-build-system-configs" (OuterVolumeSpecName: "build-system-configs") pod "2a08a655-1dec-478f-ac65-ef93056e82bb" (UID: "2a08a655-1dec-478f-ac65-ef93056e82bb"). InnerVolumeSpecName "build-system-configs". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:31:03 crc kubenswrapper[4953]: I1203 00:31:03.237799 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2a08a655-1dec-478f-ac65-ef93056e82bb-build-ca-bundles" (OuterVolumeSpecName: "build-ca-bundles") pod "2a08a655-1dec-478f-ac65-ef93056e82bb" (UID: "2a08a655-1dec-478f-ac65-ef93056e82bb"). InnerVolumeSpecName "build-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:31:03 crc kubenswrapper[4953]: I1203 00:31:03.238284 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"builder-dockercfg-9prwd-push\" (UniqueName: \"kubernetes.io/secret/2a08a655-1dec-478f-ac65-ef93056e82bb-builder-dockercfg-9prwd-push\") pod \"2a08a655-1dec-478f-ac65-ef93056e82bb\" (UID: \"2a08a655-1dec-478f-ac65-ef93056e82bb\") " Dec 03 00:31:03 crc kubenswrapper[4953]: I1203 00:31:03.238355 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/2a08a655-1dec-478f-ac65-ef93056e82bb-build-proxy-ca-bundles\") pod \"2a08a655-1dec-478f-ac65-ef93056e82bb\" (UID: \"2a08a655-1dec-478f-ac65-ef93056e82bb\") " Dec 03 00:31:03 crc kubenswrapper[4953]: I1203 00:31:03.238393 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"builder-dockercfg-9prwd-pull\" (UniqueName: \"kubernetes.io/secret/2a08a655-1dec-478f-ac65-ef93056e82bb-builder-dockercfg-9prwd-pull\") pod \"2a08a655-1dec-478f-ac65-ef93056e82bb\" (UID: \"2a08a655-1dec-478f-ac65-ef93056e82bb\") " Dec 03 00:31:03 crc kubenswrapper[4953]: I1203 00:31:03.238289 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2a08a655-1dec-478f-ac65-ef93056e82bb-container-storage-run" (OuterVolumeSpecName: "container-storage-run") pod "2a08a655-1dec-478f-ac65-ef93056e82bb" (UID: "2a08a655-1dec-478f-ac65-ef93056e82bb"). InnerVolumeSpecName "container-storage-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 00:31:03 crc kubenswrapper[4953]: I1203 00:31:03.239091 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2a08a655-1dec-478f-ac65-ef93056e82bb-build-proxy-ca-bundles" (OuterVolumeSpecName: "build-proxy-ca-bundles") pod "2a08a655-1dec-478f-ac65-ef93056e82bb" (UID: "2a08a655-1dec-478f-ac65-ef93056e82bb"). InnerVolumeSpecName "build-proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:31:03 crc kubenswrapper[4953]: I1203 00:31:03.239101 4953 reconciler_common.go:293] "Volume detached for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/2a08a655-1dec-478f-ac65-ef93056e82bb-build-ca-bundles\") on node \"crc\" DevicePath \"\"" Dec 03 00:31:03 crc kubenswrapper[4953]: I1203 00:31:03.239192 4953 reconciler_common.go:293] "Volume detached for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/2a08a655-1dec-478f-ac65-ef93056e82bb-buildcachedir\") on node \"crc\" DevicePath \"\"" Dec 03 00:31:03 crc kubenswrapper[4953]: I1203 00:31:03.239210 4953 reconciler_common.go:293] "Volume detached for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/2a08a655-1dec-478f-ac65-ef93056e82bb-container-storage-run\") on node \"crc\" DevicePath \"\"" Dec 03 00:31:03 crc kubenswrapper[4953]: I1203 00:31:03.239227 4953 reconciler_common.go:293] "Volume detached for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/2a08a655-1dec-478f-ac65-ef93056e82bb-node-pullsecrets\") on node \"crc\" DevicePath \"\"" Dec 03 00:31:03 crc kubenswrapper[4953]: I1203 00:31:03.239240 4953 reconciler_common.go:293] "Volume detached for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/2a08a655-1dec-478f-ac65-ef93056e82bb-build-system-configs\") on node \"crc\" DevicePath \"\"" Dec 03 00:31:03 crc kubenswrapper[4953]: I1203 00:31:03.246815 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2a08a655-1dec-478f-ac65-ef93056e82bb-kube-api-access-2bf8k" (OuterVolumeSpecName: "kube-api-access-2bf8k") pod "2a08a655-1dec-478f-ac65-ef93056e82bb" (UID: "2a08a655-1dec-478f-ac65-ef93056e82bb"). InnerVolumeSpecName "kube-api-access-2bf8k". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:31:03 crc kubenswrapper[4953]: I1203 00:31:03.248036 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2a08a655-1dec-478f-ac65-ef93056e82bb-builder-dockercfg-9prwd-pull" (OuterVolumeSpecName: "builder-dockercfg-9prwd-pull") pod "2a08a655-1dec-478f-ac65-ef93056e82bb" (UID: "2a08a655-1dec-478f-ac65-ef93056e82bb"). InnerVolumeSpecName "builder-dockercfg-9prwd-pull". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:31:03 crc kubenswrapper[4953]: I1203 00:31:03.249781 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2a08a655-1dec-478f-ac65-ef93056e82bb-builder-dockercfg-9prwd-push" (OuterVolumeSpecName: "builder-dockercfg-9prwd-push") pod "2a08a655-1dec-478f-ac65-ef93056e82bb" (UID: "2a08a655-1dec-478f-ac65-ef93056e82bb"). InnerVolumeSpecName "builder-dockercfg-9prwd-push". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:31:03 crc kubenswrapper[4953]: I1203 00:31:03.251541 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2a08a655-1dec-478f-ac65-ef93056e82bb-buildworkdir" (OuterVolumeSpecName: "buildworkdir") pod "2a08a655-1dec-478f-ac65-ef93056e82bb" (UID: "2a08a655-1dec-478f-ac65-ef93056e82bb"). InnerVolumeSpecName "buildworkdir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 00:31:03 crc kubenswrapper[4953]: I1203 00:31:03.340448 4953 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2bf8k\" (UniqueName: \"kubernetes.io/projected/2a08a655-1dec-478f-ac65-ef93056e82bb-kube-api-access-2bf8k\") on node \"crc\" DevicePath \"\"" Dec 03 00:31:03 crc kubenswrapper[4953]: I1203 00:31:03.340484 4953 reconciler_common.go:293] "Volume detached for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/2a08a655-1dec-478f-ac65-ef93056e82bb-build-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Dec 03 00:31:03 crc kubenswrapper[4953]: I1203 00:31:03.340496 4953 reconciler_common.go:293] "Volume detached for volume \"builder-dockercfg-9prwd-push\" (UniqueName: \"kubernetes.io/secret/2a08a655-1dec-478f-ac65-ef93056e82bb-builder-dockercfg-9prwd-push\") on node \"crc\" DevicePath \"\"" Dec 03 00:31:03 crc kubenswrapper[4953]: I1203 00:31:03.340508 4953 reconciler_common.go:293] "Volume detached for volume \"builder-dockercfg-9prwd-pull\" (UniqueName: \"kubernetes.io/secret/2a08a655-1dec-478f-ac65-ef93056e82bb-builder-dockercfg-9prwd-pull\") on node \"crc\" DevicePath \"\"" Dec 03 00:31:03 crc kubenswrapper[4953]: I1203 00:31:03.340517 4953 reconciler_common.go:293] "Volume detached for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/2a08a655-1dec-478f-ac65-ef93056e82bb-buildworkdir\") on node \"crc\" DevicePath \"\"" Dec 03 00:31:03 crc kubenswrapper[4953]: I1203 00:31:03.564256 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2a08a655-1dec-478f-ac65-ef93056e82bb-build-blob-cache" (OuterVolumeSpecName: "build-blob-cache") pod "2a08a655-1dec-478f-ac65-ef93056e82bb" (UID: "2a08a655-1dec-478f-ac65-ef93056e82bb"). InnerVolumeSpecName "build-blob-cache". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 00:31:03 crc kubenswrapper[4953]: I1203 00:31:03.645239 4953 reconciler_common.go:293] "Volume detached for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/2a08a655-1dec-478f-ac65-ef93056e82bb-build-blob-cache\") on node \"crc\" DevicePath \"\"" Dec 03 00:31:03 crc kubenswrapper[4953]: I1203 00:31:03.878286 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/sg-core-2-build" event={"ID":"2a08a655-1dec-478f-ac65-ef93056e82bb","Type":"ContainerDied","Data":"23d5a199350d200228f309ef2bcc1454f9d8956d2234a2d6ca304b76cc029e38"} Dec 03 00:31:03 crc kubenswrapper[4953]: I1203 00:31:03.878347 4953 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="23d5a199350d200228f309ef2bcc1454f9d8956d2234a2d6ca304b76cc029e38" Dec 03 00:31:03 crc kubenswrapper[4953]: I1203 00:31:03.878452 4953 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/sg-core-2-build" Dec 03 00:31:05 crc kubenswrapper[4953]: I1203 00:31:05.882747 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2a08a655-1dec-478f-ac65-ef93056e82bb-container-storage-root" (OuterVolumeSpecName: "container-storage-root") pod "2a08a655-1dec-478f-ac65-ef93056e82bb" (UID: "2a08a655-1dec-478f-ac65-ef93056e82bb"). InnerVolumeSpecName "container-storage-root". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 00:31:05 crc kubenswrapper[4953]: I1203 00:31:05.983788 4953 reconciler_common.go:293] "Volume detached for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/2a08a655-1dec-478f-ac65-ef93056e82bb-container-storage-root\") on node \"crc\" DevicePath \"\"" Dec 03 00:31:07 crc kubenswrapper[4953]: I1203 00:31:07.981745 4953 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["service-telemetry/sg-bridge-1-build"] Dec 03 00:31:07 crc kubenswrapper[4953]: E1203 00:31:07.982148 4953 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2a08a655-1dec-478f-ac65-ef93056e82bb" containerName="manage-dockerfile" Dec 03 00:31:07 crc kubenswrapper[4953]: I1203 00:31:07.982167 4953 state_mem.go:107] "Deleted CPUSet assignment" podUID="2a08a655-1dec-478f-ac65-ef93056e82bb" containerName="manage-dockerfile" Dec 03 00:31:07 crc kubenswrapper[4953]: E1203 00:31:07.982178 4953 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2a08a655-1dec-478f-ac65-ef93056e82bb" containerName="git-clone" Dec 03 00:31:07 crc kubenswrapper[4953]: I1203 00:31:07.982187 4953 state_mem.go:107] "Deleted CPUSet assignment" podUID="2a08a655-1dec-478f-ac65-ef93056e82bb" containerName="git-clone" Dec 03 00:31:07 crc kubenswrapper[4953]: E1203 00:31:07.982208 4953 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dad9c3d8-bc04-466b-9d82-4120eb7e8139" containerName="collect-profiles" Dec 03 00:31:07 crc kubenswrapper[4953]: I1203 00:31:07.982215 4953 state_mem.go:107] "Deleted CPUSet assignment" podUID="dad9c3d8-bc04-466b-9d82-4120eb7e8139" containerName="collect-profiles" Dec 03 00:31:07 crc kubenswrapper[4953]: E1203 00:31:07.982234 4953 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2a08a655-1dec-478f-ac65-ef93056e82bb" containerName="docker-build" Dec 03 00:31:07 crc kubenswrapper[4953]: I1203 00:31:07.982243 4953 state_mem.go:107] "Deleted CPUSet assignment" podUID="2a08a655-1dec-478f-ac65-ef93056e82bb" containerName="docker-build" Dec 03 00:31:07 crc kubenswrapper[4953]: I1203 00:31:07.982416 4953 memory_manager.go:354] "RemoveStaleState removing state" podUID="dad9c3d8-bc04-466b-9d82-4120eb7e8139" containerName="collect-profiles" Dec 03 00:31:07 crc kubenswrapper[4953]: I1203 00:31:07.982430 4953 memory_manager.go:354] "RemoveStaleState removing state" podUID="2a08a655-1dec-478f-ac65-ef93056e82bb" containerName="docker-build" Dec 03 00:31:07 crc kubenswrapper[4953]: I1203 00:31:07.983675 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/sg-bridge-1-build" Dec 03 00:31:07 crc kubenswrapper[4953]: I1203 00:31:07.987128 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"sg-bridge-1-global-ca" Dec 03 00:31:07 crc kubenswrapper[4953]: I1203 00:31:07.987173 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"sg-bridge-1-ca" Dec 03 00:31:07 crc kubenswrapper[4953]: I1203 00:31:07.987231 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"sg-bridge-1-sys-config" Dec 03 00:31:07 crc kubenswrapper[4953]: I1203 00:31:07.988453 4953 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"builder-dockercfg-9prwd" Dec 03 00:31:08 crc kubenswrapper[4953]: I1203 00:31:08.001879 4953 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/sg-bridge-1-build"] Dec 03 00:31:08 crc kubenswrapper[4953]: I1203 00:31:08.114109 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/a9a989b9-cf8d-4baf-89ed-2b90c20af718-build-proxy-ca-bundles\") pod \"sg-bridge-1-build\" (UID: \"a9a989b9-cf8d-4baf-89ed-2b90c20af718\") " pod="service-telemetry/sg-bridge-1-build" Dec 03 00:31:08 crc kubenswrapper[4953]: I1203 00:31:08.114177 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/a9a989b9-cf8d-4baf-89ed-2b90c20af718-build-system-configs\") pod \"sg-bridge-1-build\" (UID: \"a9a989b9-cf8d-4baf-89ed-2b90c20af718\") " pod="service-telemetry/sg-bridge-1-build" Dec 03 00:31:08 crc kubenswrapper[4953]: I1203 00:31:08.114203 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"builder-dockercfg-9prwd-push\" (UniqueName: \"kubernetes.io/secret/a9a989b9-cf8d-4baf-89ed-2b90c20af718-builder-dockercfg-9prwd-push\") pod \"sg-bridge-1-build\" (UID: \"a9a989b9-cf8d-4baf-89ed-2b90c20af718\") " pod="service-telemetry/sg-bridge-1-build" Dec 03 00:31:08 crc kubenswrapper[4953]: I1203 00:31:08.114235 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/a9a989b9-cf8d-4baf-89ed-2b90c20af718-buildworkdir\") pod \"sg-bridge-1-build\" (UID: \"a9a989b9-cf8d-4baf-89ed-2b90c20af718\") " pod="service-telemetry/sg-bridge-1-build" Dec 03 00:31:08 crc kubenswrapper[4953]: I1203 00:31:08.114293 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/a9a989b9-cf8d-4baf-89ed-2b90c20af718-node-pullsecrets\") pod \"sg-bridge-1-build\" (UID: \"a9a989b9-cf8d-4baf-89ed-2b90c20af718\") " pod="service-telemetry/sg-bridge-1-build" Dec 03 00:31:08 crc kubenswrapper[4953]: I1203 00:31:08.114337 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/a9a989b9-cf8d-4baf-89ed-2b90c20af718-container-storage-run\") pod \"sg-bridge-1-build\" (UID: \"a9a989b9-cf8d-4baf-89ed-2b90c20af718\") " pod="service-telemetry/sg-bridge-1-build" Dec 03 00:31:08 crc kubenswrapper[4953]: I1203 00:31:08.114376 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4xj64\" (UniqueName: \"kubernetes.io/projected/a9a989b9-cf8d-4baf-89ed-2b90c20af718-kube-api-access-4xj64\") pod \"sg-bridge-1-build\" (UID: \"a9a989b9-cf8d-4baf-89ed-2b90c20af718\") " pod="service-telemetry/sg-bridge-1-build" Dec 03 00:31:08 crc kubenswrapper[4953]: I1203 00:31:08.114477 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/a9a989b9-cf8d-4baf-89ed-2b90c20af718-container-storage-root\") pod \"sg-bridge-1-build\" (UID: \"a9a989b9-cf8d-4baf-89ed-2b90c20af718\") " pod="service-telemetry/sg-bridge-1-build" Dec 03 00:31:08 crc kubenswrapper[4953]: I1203 00:31:08.114501 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"builder-dockercfg-9prwd-pull\" (UniqueName: \"kubernetes.io/secret/a9a989b9-cf8d-4baf-89ed-2b90c20af718-builder-dockercfg-9prwd-pull\") pod \"sg-bridge-1-build\" (UID: \"a9a989b9-cf8d-4baf-89ed-2b90c20af718\") " pod="service-telemetry/sg-bridge-1-build" Dec 03 00:31:08 crc kubenswrapper[4953]: I1203 00:31:08.114660 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/a9a989b9-cf8d-4baf-89ed-2b90c20af718-buildcachedir\") pod \"sg-bridge-1-build\" (UID: \"a9a989b9-cf8d-4baf-89ed-2b90c20af718\") " pod="service-telemetry/sg-bridge-1-build" Dec 03 00:31:08 crc kubenswrapper[4953]: I1203 00:31:08.114777 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/a9a989b9-cf8d-4baf-89ed-2b90c20af718-build-ca-bundles\") pod \"sg-bridge-1-build\" (UID: \"a9a989b9-cf8d-4baf-89ed-2b90c20af718\") " pod="service-telemetry/sg-bridge-1-build" Dec 03 00:31:08 crc kubenswrapper[4953]: I1203 00:31:08.114849 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/a9a989b9-cf8d-4baf-89ed-2b90c20af718-build-blob-cache\") pod \"sg-bridge-1-build\" (UID: \"a9a989b9-cf8d-4baf-89ed-2b90c20af718\") " pod="service-telemetry/sg-bridge-1-build" Dec 03 00:31:08 crc kubenswrapper[4953]: I1203 00:31:08.216525 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/a9a989b9-cf8d-4baf-89ed-2b90c20af718-build-system-configs\") pod \"sg-bridge-1-build\" (UID: \"a9a989b9-cf8d-4baf-89ed-2b90c20af718\") " pod="service-telemetry/sg-bridge-1-build" Dec 03 00:31:08 crc kubenswrapper[4953]: I1203 00:31:08.216616 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"builder-dockercfg-9prwd-push\" (UniqueName: \"kubernetes.io/secret/a9a989b9-cf8d-4baf-89ed-2b90c20af718-builder-dockercfg-9prwd-push\") pod \"sg-bridge-1-build\" (UID: \"a9a989b9-cf8d-4baf-89ed-2b90c20af718\") " pod="service-telemetry/sg-bridge-1-build" Dec 03 00:31:08 crc kubenswrapper[4953]: I1203 00:31:08.216652 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/a9a989b9-cf8d-4baf-89ed-2b90c20af718-buildworkdir\") pod \"sg-bridge-1-build\" (UID: \"a9a989b9-cf8d-4baf-89ed-2b90c20af718\") " pod="service-telemetry/sg-bridge-1-build" Dec 03 00:31:08 crc kubenswrapper[4953]: I1203 00:31:08.216750 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/a9a989b9-cf8d-4baf-89ed-2b90c20af718-node-pullsecrets\") pod \"sg-bridge-1-build\" (UID: \"a9a989b9-cf8d-4baf-89ed-2b90c20af718\") " pod="service-telemetry/sg-bridge-1-build" Dec 03 00:31:08 crc kubenswrapper[4953]: I1203 00:31:08.216772 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/a9a989b9-cf8d-4baf-89ed-2b90c20af718-container-storage-run\") pod \"sg-bridge-1-build\" (UID: \"a9a989b9-cf8d-4baf-89ed-2b90c20af718\") " pod="service-telemetry/sg-bridge-1-build" Dec 03 00:31:08 crc kubenswrapper[4953]: I1203 00:31:08.216805 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4xj64\" (UniqueName: \"kubernetes.io/projected/a9a989b9-cf8d-4baf-89ed-2b90c20af718-kube-api-access-4xj64\") pod \"sg-bridge-1-build\" (UID: \"a9a989b9-cf8d-4baf-89ed-2b90c20af718\") " pod="service-telemetry/sg-bridge-1-build" Dec 03 00:31:08 crc kubenswrapper[4953]: I1203 00:31:08.216837 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/a9a989b9-cf8d-4baf-89ed-2b90c20af718-container-storage-root\") pod \"sg-bridge-1-build\" (UID: \"a9a989b9-cf8d-4baf-89ed-2b90c20af718\") " pod="service-telemetry/sg-bridge-1-build" Dec 03 00:31:08 crc kubenswrapper[4953]: I1203 00:31:08.216865 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"builder-dockercfg-9prwd-pull\" (UniqueName: \"kubernetes.io/secret/a9a989b9-cf8d-4baf-89ed-2b90c20af718-builder-dockercfg-9prwd-pull\") pod \"sg-bridge-1-build\" (UID: \"a9a989b9-cf8d-4baf-89ed-2b90c20af718\") " pod="service-telemetry/sg-bridge-1-build" Dec 03 00:31:08 crc kubenswrapper[4953]: I1203 00:31:08.216889 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/a9a989b9-cf8d-4baf-89ed-2b90c20af718-buildcachedir\") pod \"sg-bridge-1-build\" (UID: \"a9a989b9-cf8d-4baf-89ed-2b90c20af718\") " pod="service-telemetry/sg-bridge-1-build" Dec 03 00:31:08 crc kubenswrapper[4953]: I1203 00:31:08.216917 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/a9a989b9-cf8d-4baf-89ed-2b90c20af718-build-ca-bundles\") pod \"sg-bridge-1-build\" (UID: \"a9a989b9-cf8d-4baf-89ed-2b90c20af718\") " pod="service-telemetry/sg-bridge-1-build" Dec 03 00:31:08 crc kubenswrapper[4953]: I1203 00:31:08.216949 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/a9a989b9-cf8d-4baf-89ed-2b90c20af718-build-blob-cache\") pod \"sg-bridge-1-build\" (UID: \"a9a989b9-cf8d-4baf-89ed-2b90c20af718\") " pod="service-telemetry/sg-bridge-1-build" Dec 03 00:31:08 crc kubenswrapper[4953]: I1203 00:31:08.216987 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/a9a989b9-cf8d-4baf-89ed-2b90c20af718-build-proxy-ca-bundles\") pod \"sg-bridge-1-build\" (UID: \"a9a989b9-cf8d-4baf-89ed-2b90c20af718\") " pod="service-telemetry/sg-bridge-1-build" Dec 03 00:31:08 crc kubenswrapper[4953]: I1203 00:31:08.217245 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/a9a989b9-cf8d-4baf-89ed-2b90c20af718-buildcachedir\") pod \"sg-bridge-1-build\" (UID: \"a9a989b9-cf8d-4baf-89ed-2b90c20af718\") " pod="service-telemetry/sg-bridge-1-build" Dec 03 00:31:08 crc kubenswrapper[4953]: I1203 00:31:08.217415 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/a9a989b9-cf8d-4baf-89ed-2b90c20af718-node-pullsecrets\") pod \"sg-bridge-1-build\" (UID: \"a9a989b9-cf8d-4baf-89ed-2b90c20af718\") " pod="service-telemetry/sg-bridge-1-build" Dec 03 00:31:08 crc kubenswrapper[4953]: I1203 00:31:08.217477 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/a9a989b9-cf8d-4baf-89ed-2b90c20af718-buildworkdir\") pod \"sg-bridge-1-build\" (UID: \"a9a989b9-cf8d-4baf-89ed-2b90c20af718\") " pod="service-telemetry/sg-bridge-1-build" Dec 03 00:31:08 crc kubenswrapper[4953]: I1203 00:31:08.218244 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/a9a989b9-cf8d-4baf-89ed-2b90c20af718-container-storage-root\") pod \"sg-bridge-1-build\" (UID: \"a9a989b9-cf8d-4baf-89ed-2b90c20af718\") " pod="service-telemetry/sg-bridge-1-build" Dec 03 00:31:08 crc kubenswrapper[4953]: I1203 00:31:08.218432 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/a9a989b9-cf8d-4baf-89ed-2b90c20af718-build-blob-cache\") pod \"sg-bridge-1-build\" (UID: \"a9a989b9-cf8d-4baf-89ed-2b90c20af718\") " pod="service-telemetry/sg-bridge-1-build" Dec 03 00:31:08 crc kubenswrapper[4953]: I1203 00:31:08.218548 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/a9a989b9-cf8d-4baf-89ed-2b90c20af718-build-system-configs\") pod \"sg-bridge-1-build\" (UID: \"a9a989b9-cf8d-4baf-89ed-2b90c20af718\") " pod="service-telemetry/sg-bridge-1-build" Dec 03 00:31:08 crc kubenswrapper[4953]: I1203 00:31:08.218675 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/a9a989b9-cf8d-4baf-89ed-2b90c20af718-build-proxy-ca-bundles\") pod \"sg-bridge-1-build\" (UID: \"a9a989b9-cf8d-4baf-89ed-2b90c20af718\") " pod="service-telemetry/sg-bridge-1-build" Dec 03 00:31:08 crc kubenswrapper[4953]: I1203 00:31:08.218773 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/a9a989b9-cf8d-4baf-89ed-2b90c20af718-container-storage-run\") pod \"sg-bridge-1-build\" (UID: \"a9a989b9-cf8d-4baf-89ed-2b90c20af718\") " pod="service-telemetry/sg-bridge-1-build" Dec 03 00:31:08 crc kubenswrapper[4953]: I1203 00:31:08.220011 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/a9a989b9-cf8d-4baf-89ed-2b90c20af718-build-ca-bundles\") pod \"sg-bridge-1-build\" (UID: \"a9a989b9-cf8d-4baf-89ed-2b90c20af718\") " pod="service-telemetry/sg-bridge-1-build" Dec 03 00:31:08 crc kubenswrapper[4953]: I1203 00:31:08.224613 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"builder-dockercfg-9prwd-pull\" (UniqueName: \"kubernetes.io/secret/a9a989b9-cf8d-4baf-89ed-2b90c20af718-builder-dockercfg-9prwd-pull\") pod \"sg-bridge-1-build\" (UID: \"a9a989b9-cf8d-4baf-89ed-2b90c20af718\") " pod="service-telemetry/sg-bridge-1-build" Dec 03 00:31:08 crc kubenswrapper[4953]: I1203 00:31:08.226881 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"builder-dockercfg-9prwd-push\" (UniqueName: \"kubernetes.io/secret/a9a989b9-cf8d-4baf-89ed-2b90c20af718-builder-dockercfg-9prwd-push\") pod \"sg-bridge-1-build\" (UID: \"a9a989b9-cf8d-4baf-89ed-2b90c20af718\") " pod="service-telemetry/sg-bridge-1-build" Dec 03 00:31:08 crc kubenswrapper[4953]: I1203 00:31:08.239484 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4xj64\" (UniqueName: \"kubernetes.io/projected/a9a989b9-cf8d-4baf-89ed-2b90c20af718-kube-api-access-4xj64\") pod \"sg-bridge-1-build\" (UID: \"a9a989b9-cf8d-4baf-89ed-2b90c20af718\") " pod="service-telemetry/sg-bridge-1-build" Dec 03 00:31:08 crc kubenswrapper[4953]: I1203 00:31:08.303867 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/sg-bridge-1-build" Dec 03 00:31:08 crc kubenswrapper[4953]: I1203 00:31:08.538359 4953 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/sg-bridge-1-build"] Dec 03 00:31:08 crc kubenswrapper[4953]: I1203 00:31:08.921486 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/sg-bridge-1-build" event={"ID":"a9a989b9-cf8d-4baf-89ed-2b90c20af718","Type":"ContainerStarted","Data":"3cb54897db53ff6838ae17e41a58a2d09677efbd0643e61250be41a038e74c51"} Dec 03 00:31:08 crc kubenswrapper[4953]: I1203 00:31:08.921552 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/sg-bridge-1-build" event={"ID":"a9a989b9-cf8d-4baf-89ed-2b90c20af718","Type":"ContainerStarted","Data":"f4d5f6ba16ac761aa67f0101b589df10fd4701d4193dfa9c8444d9a02fc63074"} Dec 03 00:31:09 crc kubenswrapper[4953]: I1203 00:31:09.932356 4953 generic.go:334] "Generic (PLEG): container finished" podID="a9a989b9-cf8d-4baf-89ed-2b90c20af718" containerID="3cb54897db53ff6838ae17e41a58a2d09677efbd0643e61250be41a038e74c51" exitCode=0 Dec 03 00:31:09 crc kubenswrapper[4953]: I1203 00:31:09.932446 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/sg-bridge-1-build" event={"ID":"a9a989b9-cf8d-4baf-89ed-2b90c20af718","Type":"ContainerDied","Data":"3cb54897db53ff6838ae17e41a58a2d09677efbd0643e61250be41a038e74c51"} Dec 03 00:31:10 crc kubenswrapper[4953]: I1203 00:31:10.949200 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/sg-bridge-1-build" event={"ID":"a9a989b9-cf8d-4baf-89ed-2b90c20af718","Type":"ContainerStarted","Data":"ec00e6391f6f69f43150edf6cdef76106bc797e9a0e37460a980eea02b93f843"} Dec 03 00:31:10 crc kubenswrapper[4953]: I1203 00:31:10.982435 4953 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="service-telemetry/sg-bridge-1-build" podStartSLOduration=3.982412959 podStartE2EDuration="3.982412959s" podCreationTimestamp="2025-12-03 00:31:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 00:31:10.976598656 +0000 UTC m=+1459.960879509" watchObservedRunningTime="2025-12-03 00:31:10.982412959 +0000 UTC m=+1459.966693742" Dec 03 00:31:18 crc kubenswrapper[4953]: I1203 00:31:18.136122 4953 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["service-telemetry/sg-bridge-1-build"] Dec 03 00:31:18 crc kubenswrapper[4953]: I1203 00:31:18.137149 4953 kuberuntime_container.go:808] "Killing container with a grace period" pod="service-telemetry/sg-bridge-1-build" podUID="a9a989b9-cf8d-4baf-89ed-2b90c20af718" containerName="docker-build" containerID="cri-o://ec00e6391f6f69f43150edf6cdef76106bc797e9a0e37460a980eea02b93f843" gracePeriod=30 Dec 03 00:31:18 crc kubenswrapper[4953]: I1203 00:31:18.522520 4953 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_sg-bridge-1-build_a9a989b9-cf8d-4baf-89ed-2b90c20af718/docker-build/0.log" Dec 03 00:31:18 crc kubenswrapper[4953]: I1203 00:31:18.523042 4953 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/sg-bridge-1-build" Dec 03 00:31:18 crc kubenswrapper[4953]: I1203 00:31:18.583635 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/a9a989b9-cf8d-4baf-89ed-2b90c20af718-container-storage-run\") pod \"a9a989b9-cf8d-4baf-89ed-2b90c20af718\" (UID: \"a9a989b9-cf8d-4baf-89ed-2b90c20af718\") " Dec 03 00:31:18 crc kubenswrapper[4953]: I1203 00:31:18.583685 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/a9a989b9-cf8d-4baf-89ed-2b90c20af718-container-storage-root\") pod \"a9a989b9-cf8d-4baf-89ed-2b90c20af718\" (UID: \"a9a989b9-cf8d-4baf-89ed-2b90c20af718\") " Dec 03 00:31:18 crc kubenswrapper[4953]: I1203 00:31:18.583733 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/a9a989b9-cf8d-4baf-89ed-2b90c20af718-build-blob-cache\") pod \"a9a989b9-cf8d-4baf-89ed-2b90c20af718\" (UID: \"a9a989b9-cf8d-4baf-89ed-2b90c20af718\") " Dec 03 00:31:18 crc kubenswrapper[4953]: I1203 00:31:18.583765 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/a9a989b9-cf8d-4baf-89ed-2b90c20af718-build-system-configs\") pod \"a9a989b9-cf8d-4baf-89ed-2b90c20af718\" (UID: \"a9a989b9-cf8d-4baf-89ed-2b90c20af718\") " Dec 03 00:31:18 crc kubenswrapper[4953]: I1203 00:31:18.583831 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/a9a989b9-cf8d-4baf-89ed-2b90c20af718-build-ca-bundles\") pod \"a9a989b9-cf8d-4baf-89ed-2b90c20af718\" (UID: \"a9a989b9-cf8d-4baf-89ed-2b90c20af718\") " Dec 03 00:31:18 crc kubenswrapper[4953]: I1203 00:31:18.583860 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/a9a989b9-cf8d-4baf-89ed-2b90c20af718-buildcachedir\") pod \"a9a989b9-cf8d-4baf-89ed-2b90c20af718\" (UID: \"a9a989b9-cf8d-4baf-89ed-2b90c20af718\") " Dec 03 00:31:18 crc kubenswrapper[4953]: I1203 00:31:18.583880 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/a9a989b9-cf8d-4baf-89ed-2b90c20af718-node-pullsecrets\") pod \"a9a989b9-cf8d-4baf-89ed-2b90c20af718\" (UID: \"a9a989b9-cf8d-4baf-89ed-2b90c20af718\") " Dec 03 00:31:18 crc kubenswrapper[4953]: I1203 00:31:18.583940 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"builder-dockercfg-9prwd-pull\" (UniqueName: \"kubernetes.io/secret/a9a989b9-cf8d-4baf-89ed-2b90c20af718-builder-dockercfg-9prwd-pull\") pod \"a9a989b9-cf8d-4baf-89ed-2b90c20af718\" (UID: \"a9a989b9-cf8d-4baf-89ed-2b90c20af718\") " Dec 03 00:31:18 crc kubenswrapper[4953]: I1203 00:31:18.583992 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/a9a989b9-cf8d-4baf-89ed-2b90c20af718-build-proxy-ca-bundles\") pod \"a9a989b9-cf8d-4baf-89ed-2b90c20af718\" (UID: \"a9a989b9-cf8d-4baf-89ed-2b90c20af718\") " Dec 03 00:31:18 crc kubenswrapper[4953]: I1203 00:31:18.584019 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"builder-dockercfg-9prwd-push\" (UniqueName: \"kubernetes.io/secret/a9a989b9-cf8d-4baf-89ed-2b90c20af718-builder-dockercfg-9prwd-push\") pod \"a9a989b9-cf8d-4baf-89ed-2b90c20af718\" (UID: \"a9a989b9-cf8d-4baf-89ed-2b90c20af718\") " Dec 03 00:31:18 crc kubenswrapper[4953]: I1203 00:31:18.584069 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/a9a989b9-cf8d-4baf-89ed-2b90c20af718-buildworkdir\") pod \"a9a989b9-cf8d-4baf-89ed-2b90c20af718\" (UID: \"a9a989b9-cf8d-4baf-89ed-2b90c20af718\") " Dec 03 00:31:18 crc kubenswrapper[4953]: I1203 00:31:18.584093 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4xj64\" (UniqueName: \"kubernetes.io/projected/a9a989b9-cf8d-4baf-89ed-2b90c20af718-kube-api-access-4xj64\") pod \"a9a989b9-cf8d-4baf-89ed-2b90c20af718\" (UID: \"a9a989b9-cf8d-4baf-89ed-2b90c20af718\") " Dec 03 00:31:18 crc kubenswrapper[4953]: I1203 00:31:18.585060 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a9a989b9-cf8d-4baf-89ed-2b90c20af718-node-pullsecrets" (OuterVolumeSpecName: "node-pullsecrets") pod "a9a989b9-cf8d-4baf-89ed-2b90c20af718" (UID: "a9a989b9-cf8d-4baf-89ed-2b90c20af718"). InnerVolumeSpecName "node-pullsecrets". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 00:31:18 crc kubenswrapper[4953]: I1203 00:31:18.585215 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a9a989b9-cf8d-4baf-89ed-2b90c20af718-buildcachedir" (OuterVolumeSpecName: "buildcachedir") pod "a9a989b9-cf8d-4baf-89ed-2b90c20af718" (UID: "a9a989b9-cf8d-4baf-89ed-2b90c20af718"). InnerVolumeSpecName "buildcachedir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 00:31:18 crc kubenswrapper[4953]: I1203 00:31:18.585792 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a9a989b9-cf8d-4baf-89ed-2b90c20af718-build-proxy-ca-bundles" (OuterVolumeSpecName: "build-proxy-ca-bundles") pod "a9a989b9-cf8d-4baf-89ed-2b90c20af718" (UID: "a9a989b9-cf8d-4baf-89ed-2b90c20af718"). InnerVolumeSpecName "build-proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:31:18 crc kubenswrapper[4953]: I1203 00:31:18.585838 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a9a989b9-cf8d-4baf-89ed-2b90c20af718-build-ca-bundles" (OuterVolumeSpecName: "build-ca-bundles") pod "a9a989b9-cf8d-4baf-89ed-2b90c20af718" (UID: "a9a989b9-cf8d-4baf-89ed-2b90c20af718"). InnerVolumeSpecName "build-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:31:18 crc kubenswrapper[4953]: I1203 00:31:18.586110 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a9a989b9-cf8d-4baf-89ed-2b90c20af718-build-system-configs" (OuterVolumeSpecName: "build-system-configs") pod "a9a989b9-cf8d-4baf-89ed-2b90c20af718" (UID: "a9a989b9-cf8d-4baf-89ed-2b90c20af718"). InnerVolumeSpecName "build-system-configs". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:31:18 crc kubenswrapper[4953]: I1203 00:31:18.586160 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a9a989b9-cf8d-4baf-89ed-2b90c20af718-container-storage-run" (OuterVolumeSpecName: "container-storage-run") pod "a9a989b9-cf8d-4baf-89ed-2b90c20af718" (UID: "a9a989b9-cf8d-4baf-89ed-2b90c20af718"). InnerVolumeSpecName "container-storage-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 00:31:18 crc kubenswrapper[4953]: I1203 00:31:18.587662 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a9a989b9-cf8d-4baf-89ed-2b90c20af718-buildworkdir" (OuterVolumeSpecName: "buildworkdir") pod "a9a989b9-cf8d-4baf-89ed-2b90c20af718" (UID: "a9a989b9-cf8d-4baf-89ed-2b90c20af718"). InnerVolumeSpecName "buildworkdir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 00:31:18 crc kubenswrapper[4953]: I1203 00:31:18.591864 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a9a989b9-cf8d-4baf-89ed-2b90c20af718-builder-dockercfg-9prwd-pull" (OuterVolumeSpecName: "builder-dockercfg-9prwd-pull") pod "a9a989b9-cf8d-4baf-89ed-2b90c20af718" (UID: "a9a989b9-cf8d-4baf-89ed-2b90c20af718"). InnerVolumeSpecName "builder-dockercfg-9prwd-pull". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:31:18 crc kubenswrapper[4953]: I1203 00:31:18.592137 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a9a989b9-cf8d-4baf-89ed-2b90c20af718-kube-api-access-4xj64" (OuterVolumeSpecName: "kube-api-access-4xj64") pod "a9a989b9-cf8d-4baf-89ed-2b90c20af718" (UID: "a9a989b9-cf8d-4baf-89ed-2b90c20af718"). InnerVolumeSpecName "kube-api-access-4xj64". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:31:18 crc kubenswrapper[4953]: I1203 00:31:18.593089 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a9a989b9-cf8d-4baf-89ed-2b90c20af718-builder-dockercfg-9prwd-push" (OuterVolumeSpecName: "builder-dockercfg-9prwd-push") pod "a9a989b9-cf8d-4baf-89ed-2b90c20af718" (UID: "a9a989b9-cf8d-4baf-89ed-2b90c20af718"). InnerVolumeSpecName "builder-dockercfg-9prwd-push". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:31:18 crc kubenswrapper[4953]: I1203 00:31:18.675762 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a9a989b9-cf8d-4baf-89ed-2b90c20af718-build-blob-cache" (OuterVolumeSpecName: "build-blob-cache") pod "a9a989b9-cf8d-4baf-89ed-2b90c20af718" (UID: "a9a989b9-cf8d-4baf-89ed-2b90c20af718"). InnerVolumeSpecName "build-blob-cache". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 00:31:18 crc kubenswrapper[4953]: I1203 00:31:18.686878 4953 reconciler_common.go:293] "Volume detached for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/a9a989b9-cf8d-4baf-89ed-2b90c20af718-buildworkdir\") on node \"crc\" DevicePath \"\"" Dec 03 00:31:18 crc kubenswrapper[4953]: I1203 00:31:18.686915 4953 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4xj64\" (UniqueName: \"kubernetes.io/projected/a9a989b9-cf8d-4baf-89ed-2b90c20af718-kube-api-access-4xj64\") on node \"crc\" DevicePath \"\"" Dec 03 00:31:18 crc kubenswrapper[4953]: I1203 00:31:18.686936 4953 reconciler_common.go:293] "Volume detached for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/a9a989b9-cf8d-4baf-89ed-2b90c20af718-container-storage-run\") on node \"crc\" DevicePath \"\"" Dec 03 00:31:18 crc kubenswrapper[4953]: I1203 00:31:18.686951 4953 reconciler_common.go:293] "Volume detached for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/a9a989b9-cf8d-4baf-89ed-2b90c20af718-build-blob-cache\") on node \"crc\" DevicePath \"\"" Dec 03 00:31:18 crc kubenswrapper[4953]: I1203 00:31:18.686965 4953 reconciler_common.go:293] "Volume detached for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/a9a989b9-cf8d-4baf-89ed-2b90c20af718-build-system-configs\") on node \"crc\" DevicePath \"\"" Dec 03 00:31:18 crc kubenswrapper[4953]: I1203 00:31:18.686992 4953 reconciler_common.go:293] "Volume detached for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/a9a989b9-cf8d-4baf-89ed-2b90c20af718-build-ca-bundles\") on node \"crc\" DevicePath \"\"" Dec 03 00:31:18 crc kubenswrapper[4953]: I1203 00:31:18.687000 4953 reconciler_common.go:293] "Volume detached for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/a9a989b9-cf8d-4baf-89ed-2b90c20af718-buildcachedir\") on node \"crc\" DevicePath \"\"" Dec 03 00:31:18 crc kubenswrapper[4953]: I1203 00:31:18.687009 4953 reconciler_common.go:293] "Volume detached for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/a9a989b9-cf8d-4baf-89ed-2b90c20af718-node-pullsecrets\") on node \"crc\" DevicePath \"\"" Dec 03 00:31:18 crc kubenswrapper[4953]: I1203 00:31:18.687019 4953 reconciler_common.go:293] "Volume detached for volume \"builder-dockercfg-9prwd-pull\" (UniqueName: \"kubernetes.io/secret/a9a989b9-cf8d-4baf-89ed-2b90c20af718-builder-dockercfg-9prwd-pull\") on node \"crc\" DevicePath \"\"" Dec 03 00:31:18 crc kubenswrapper[4953]: I1203 00:31:18.687032 4953 reconciler_common.go:293] "Volume detached for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/a9a989b9-cf8d-4baf-89ed-2b90c20af718-build-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Dec 03 00:31:18 crc kubenswrapper[4953]: I1203 00:31:18.687043 4953 reconciler_common.go:293] "Volume detached for volume \"builder-dockercfg-9prwd-push\" (UniqueName: \"kubernetes.io/secret/a9a989b9-cf8d-4baf-89ed-2b90c20af718-builder-dockercfg-9prwd-push\") on node \"crc\" DevicePath \"\"" Dec 03 00:31:18 crc kubenswrapper[4953]: I1203 00:31:18.984628 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a9a989b9-cf8d-4baf-89ed-2b90c20af718-container-storage-root" (OuterVolumeSpecName: "container-storage-root") pod "a9a989b9-cf8d-4baf-89ed-2b90c20af718" (UID: "a9a989b9-cf8d-4baf-89ed-2b90c20af718"). InnerVolumeSpecName "container-storage-root". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 00:31:18 crc kubenswrapper[4953]: I1203 00:31:18.992233 4953 reconciler_common.go:293] "Volume detached for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/a9a989b9-cf8d-4baf-89ed-2b90c20af718-container-storage-root\") on node \"crc\" DevicePath \"\"" Dec 03 00:31:19 crc kubenswrapper[4953]: I1203 00:31:19.004206 4953 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_sg-bridge-1-build_a9a989b9-cf8d-4baf-89ed-2b90c20af718/docker-build/0.log" Dec 03 00:31:19 crc kubenswrapper[4953]: I1203 00:31:19.004565 4953 generic.go:334] "Generic (PLEG): container finished" podID="a9a989b9-cf8d-4baf-89ed-2b90c20af718" containerID="ec00e6391f6f69f43150edf6cdef76106bc797e9a0e37460a980eea02b93f843" exitCode=1 Dec 03 00:31:19 crc kubenswrapper[4953]: I1203 00:31:19.004644 4953 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/sg-bridge-1-build" Dec 03 00:31:19 crc kubenswrapper[4953]: I1203 00:31:19.004633 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/sg-bridge-1-build" event={"ID":"a9a989b9-cf8d-4baf-89ed-2b90c20af718","Type":"ContainerDied","Data":"ec00e6391f6f69f43150edf6cdef76106bc797e9a0e37460a980eea02b93f843"} Dec 03 00:31:19 crc kubenswrapper[4953]: I1203 00:31:19.004782 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/sg-bridge-1-build" event={"ID":"a9a989b9-cf8d-4baf-89ed-2b90c20af718","Type":"ContainerDied","Data":"f4d5f6ba16ac761aa67f0101b589df10fd4701d4193dfa9c8444d9a02fc63074"} Dec 03 00:31:19 crc kubenswrapper[4953]: I1203 00:31:19.004816 4953 scope.go:117] "RemoveContainer" containerID="ec00e6391f6f69f43150edf6cdef76106bc797e9a0e37460a980eea02b93f843" Dec 03 00:31:19 crc kubenswrapper[4953]: I1203 00:31:19.051805 4953 scope.go:117] "RemoveContainer" containerID="3cb54897db53ff6838ae17e41a58a2d09677efbd0643e61250be41a038e74c51" Dec 03 00:31:19 crc kubenswrapper[4953]: I1203 00:31:19.058780 4953 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["service-telemetry/sg-bridge-1-build"] Dec 03 00:31:19 crc kubenswrapper[4953]: I1203 00:31:19.065845 4953 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["service-telemetry/sg-bridge-1-build"] Dec 03 00:31:19 crc kubenswrapper[4953]: I1203 00:31:19.090844 4953 scope.go:117] "RemoveContainer" containerID="ec00e6391f6f69f43150edf6cdef76106bc797e9a0e37460a980eea02b93f843" Dec 03 00:31:19 crc kubenswrapper[4953]: E1203 00:31:19.091698 4953 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ec00e6391f6f69f43150edf6cdef76106bc797e9a0e37460a980eea02b93f843\": container with ID starting with ec00e6391f6f69f43150edf6cdef76106bc797e9a0e37460a980eea02b93f843 not found: ID does not exist" containerID="ec00e6391f6f69f43150edf6cdef76106bc797e9a0e37460a980eea02b93f843" Dec 03 00:31:19 crc kubenswrapper[4953]: I1203 00:31:19.091765 4953 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ec00e6391f6f69f43150edf6cdef76106bc797e9a0e37460a980eea02b93f843"} err="failed to get container status \"ec00e6391f6f69f43150edf6cdef76106bc797e9a0e37460a980eea02b93f843\": rpc error: code = NotFound desc = could not find container \"ec00e6391f6f69f43150edf6cdef76106bc797e9a0e37460a980eea02b93f843\": container with ID starting with ec00e6391f6f69f43150edf6cdef76106bc797e9a0e37460a980eea02b93f843 not found: ID does not exist" Dec 03 00:31:19 crc kubenswrapper[4953]: I1203 00:31:19.091810 4953 scope.go:117] "RemoveContainer" containerID="3cb54897db53ff6838ae17e41a58a2d09677efbd0643e61250be41a038e74c51" Dec 03 00:31:19 crc kubenswrapper[4953]: E1203 00:31:19.093345 4953 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3cb54897db53ff6838ae17e41a58a2d09677efbd0643e61250be41a038e74c51\": container with ID starting with 3cb54897db53ff6838ae17e41a58a2d09677efbd0643e61250be41a038e74c51 not found: ID does not exist" containerID="3cb54897db53ff6838ae17e41a58a2d09677efbd0643e61250be41a038e74c51" Dec 03 00:31:19 crc kubenswrapper[4953]: I1203 00:31:19.093376 4953 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3cb54897db53ff6838ae17e41a58a2d09677efbd0643e61250be41a038e74c51"} err="failed to get container status \"3cb54897db53ff6838ae17e41a58a2d09677efbd0643e61250be41a038e74c51\": rpc error: code = NotFound desc = could not find container \"3cb54897db53ff6838ae17e41a58a2d09677efbd0643e61250be41a038e74c51\": container with ID starting with 3cb54897db53ff6838ae17e41a58a2d09677efbd0643e61250be41a038e74c51 not found: ID does not exist" Dec 03 00:31:19 crc kubenswrapper[4953]: I1203 00:31:19.356725 4953 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a9a989b9-cf8d-4baf-89ed-2b90c20af718" path="/var/lib/kubelet/pods/a9a989b9-cf8d-4baf-89ed-2b90c20af718/volumes" Dec 03 00:31:19 crc kubenswrapper[4953]: I1203 00:31:19.812057 4953 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["service-telemetry/sg-bridge-2-build"] Dec 03 00:31:19 crc kubenswrapper[4953]: E1203 00:31:19.812430 4953 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a9a989b9-cf8d-4baf-89ed-2b90c20af718" containerName="manage-dockerfile" Dec 03 00:31:19 crc kubenswrapper[4953]: I1203 00:31:19.812459 4953 state_mem.go:107] "Deleted CPUSet assignment" podUID="a9a989b9-cf8d-4baf-89ed-2b90c20af718" containerName="manage-dockerfile" Dec 03 00:31:19 crc kubenswrapper[4953]: E1203 00:31:19.812490 4953 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a9a989b9-cf8d-4baf-89ed-2b90c20af718" containerName="docker-build" Dec 03 00:31:19 crc kubenswrapper[4953]: I1203 00:31:19.812504 4953 state_mem.go:107] "Deleted CPUSet assignment" podUID="a9a989b9-cf8d-4baf-89ed-2b90c20af718" containerName="docker-build" Dec 03 00:31:19 crc kubenswrapper[4953]: I1203 00:31:19.812697 4953 memory_manager.go:354] "RemoveStaleState removing state" podUID="a9a989b9-cf8d-4baf-89ed-2b90c20af718" containerName="docker-build" Dec 03 00:31:19 crc kubenswrapper[4953]: I1203 00:31:19.814208 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/sg-bridge-2-build" Dec 03 00:31:19 crc kubenswrapper[4953]: I1203 00:31:19.818560 4953 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"builder-dockercfg-9prwd" Dec 03 00:31:19 crc kubenswrapper[4953]: I1203 00:31:19.819109 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"sg-bridge-2-ca" Dec 03 00:31:19 crc kubenswrapper[4953]: I1203 00:31:19.819925 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"sg-bridge-2-global-ca" Dec 03 00:31:19 crc kubenswrapper[4953]: I1203 00:31:19.820205 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"sg-bridge-2-sys-config" Dec 03 00:31:19 crc kubenswrapper[4953]: I1203 00:31:19.846355 4953 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/sg-bridge-2-build"] Dec 03 00:31:19 crc kubenswrapper[4953]: I1203 00:31:19.909151 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/2b0995a0-793c-4aac-9d8d-bc0dd4604b29-buildworkdir\") pod \"sg-bridge-2-build\" (UID: \"2b0995a0-793c-4aac-9d8d-bc0dd4604b29\") " pod="service-telemetry/sg-bridge-2-build" Dec 03 00:31:19 crc kubenswrapper[4953]: I1203 00:31:19.909216 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/2b0995a0-793c-4aac-9d8d-bc0dd4604b29-build-system-configs\") pod \"sg-bridge-2-build\" (UID: \"2b0995a0-793c-4aac-9d8d-bc0dd4604b29\") " pod="service-telemetry/sg-bridge-2-build" Dec 03 00:31:19 crc kubenswrapper[4953]: I1203 00:31:19.909248 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bxf8m\" (UniqueName: \"kubernetes.io/projected/2b0995a0-793c-4aac-9d8d-bc0dd4604b29-kube-api-access-bxf8m\") pod \"sg-bridge-2-build\" (UID: \"2b0995a0-793c-4aac-9d8d-bc0dd4604b29\") " pod="service-telemetry/sg-bridge-2-build" Dec 03 00:31:19 crc kubenswrapper[4953]: I1203 00:31:19.909286 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/2b0995a0-793c-4aac-9d8d-bc0dd4604b29-node-pullsecrets\") pod \"sg-bridge-2-build\" (UID: \"2b0995a0-793c-4aac-9d8d-bc0dd4604b29\") " pod="service-telemetry/sg-bridge-2-build" Dec 03 00:31:19 crc kubenswrapper[4953]: I1203 00:31:19.909317 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/2b0995a0-793c-4aac-9d8d-bc0dd4604b29-container-storage-root\") pod \"sg-bridge-2-build\" (UID: \"2b0995a0-793c-4aac-9d8d-bc0dd4604b29\") " pod="service-telemetry/sg-bridge-2-build" Dec 03 00:31:19 crc kubenswrapper[4953]: I1203 00:31:19.909348 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/2b0995a0-793c-4aac-9d8d-bc0dd4604b29-buildcachedir\") pod \"sg-bridge-2-build\" (UID: \"2b0995a0-793c-4aac-9d8d-bc0dd4604b29\") " pod="service-telemetry/sg-bridge-2-build" Dec 03 00:31:19 crc kubenswrapper[4953]: I1203 00:31:19.909455 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/2b0995a0-793c-4aac-9d8d-bc0dd4604b29-build-ca-bundles\") pod \"sg-bridge-2-build\" (UID: \"2b0995a0-793c-4aac-9d8d-bc0dd4604b29\") " pod="service-telemetry/sg-bridge-2-build" Dec 03 00:31:19 crc kubenswrapper[4953]: I1203 00:31:19.909502 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"builder-dockercfg-9prwd-push\" (UniqueName: \"kubernetes.io/secret/2b0995a0-793c-4aac-9d8d-bc0dd4604b29-builder-dockercfg-9prwd-push\") pod \"sg-bridge-2-build\" (UID: \"2b0995a0-793c-4aac-9d8d-bc0dd4604b29\") " pod="service-telemetry/sg-bridge-2-build" Dec 03 00:31:19 crc kubenswrapper[4953]: I1203 00:31:19.909526 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/2b0995a0-793c-4aac-9d8d-bc0dd4604b29-build-proxy-ca-bundles\") pod \"sg-bridge-2-build\" (UID: \"2b0995a0-793c-4aac-9d8d-bc0dd4604b29\") " pod="service-telemetry/sg-bridge-2-build" Dec 03 00:31:19 crc kubenswrapper[4953]: I1203 00:31:19.909558 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/2b0995a0-793c-4aac-9d8d-bc0dd4604b29-container-storage-run\") pod \"sg-bridge-2-build\" (UID: \"2b0995a0-793c-4aac-9d8d-bc0dd4604b29\") " pod="service-telemetry/sg-bridge-2-build" Dec 03 00:31:19 crc kubenswrapper[4953]: I1203 00:31:19.909603 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"builder-dockercfg-9prwd-pull\" (UniqueName: \"kubernetes.io/secret/2b0995a0-793c-4aac-9d8d-bc0dd4604b29-builder-dockercfg-9prwd-pull\") pod \"sg-bridge-2-build\" (UID: \"2b0995a0-793c-4aac-9d8d-bc0dd4604b29\") " pod="service-telemetry/sg-bridge-2-build" Dec 03 00:31:19 crc kubenswrapper[4953]: I1203 00:31:19.909633 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/2b0995a0-793c-4aac-9d8d-bc0dd4604b29-build-blob-cache\") pod \"sg-bridge-2-build\" (UID: \"2b0995a0-793c-4aac-9d8d-bc0dd4604b29\") " pod="service-telemetry/sg-bridge-2-build" Dec 03 00:31:20 crc kubenswrapper[4953]: I1203 00:31:20.010899 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"builder-dockercfg-9prwd-pull\" (UniqueName: \"kubernetes.io/secret/2b0995a0-793c-4aac-9d8d-bc0dd4604b29-builder-dockercfg-9prwd-pull\") pod \"sg-bridge-2-build\" (UID: \"2b0995a0-793c-4aac-9d8d-bc0dd4604b29\") " pod="service-telemetry/sg-bridge-2-build" Dec 03 00:31:20 crc kubenswrapper[4953]: I1203 00:31:20.010961 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/2b0995a0-793c-4aac-9d8d-bc0dd4604b29-build-blob-cache\") pod \"sg-bridge-2-build\" (UID: \"2b0995a0-793c-4aac-9d8d-bc0dd4604b29\") " pod="service-telemetry/sg-bridge-2-build" Dec 03 00:31:20 crc kubenswrapper[4953]: I1203 00:31:20.011050 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/2b0995a0-793c-4aac-9d8d-bc0dd4604b29-buildworkdir\") pod \"sg-bridge-2-build\" (UID: \"2b0995a0-793c-4aac-9d8d-bc0dd4604b29\") " pod="service-telemetry/sg-bridge-2-build" Dec 03 00:31:20 crc kubenswrapper[4953]: I1203 00:31:20.011081 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/2b0995a0-793c-4aac-9d8d-bc0dd4604b29-build-system-configs\") pod \"sg-bridge-2-build\" (UID: \"2b0995a0-793c-4aac-9d8d-bc0dd4604b29\") " pod="service-telemetry/sg-bridge-2-build" Dec 03 00:31:20 crc kubenswrapper[4953]: I1203 00:31:20.011105 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bxf8m\" (UniqueName: \"kubernetes.io/projected/2b0995a0-793c-4aac-9d8d-bc0dd4604b29-kube-api-access-bxf8m\") pod \"sg-bridge-2-build\" (UID: \"2b0995a0-793c-4aac-9d8d-bc0dd4604b29\") " pod="service-telemetry/sg-bridge-2-build" Dec 03 00:31:20 crc kubenswrapper[4953]: I1203 00:31:20.011138 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/2b0995a0-793c-4aac-9d8d-bc0dd4604b29-node-pullsecrets\") pod \"sg-bridge-2-build\" (UID: \"2b0995a0-793c-4aac-9d8d-bc0dd4604b29\") " pod="service-telemetry/sg-bridge-2-build" Dec 03 00:31:20 crc kubenswrapper[4953]: I1203 00:31:20.011163 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/2b0995a0-793c-4aac-9d8d-bc0dd4604b29-container-storage-root\") pod \"sg-bridge-2-build\" (UID: \"2b0995a0-793c-4aac-9d8d-bc0dd4604b29\") " pod="service-telemetry/sg-bridge-2-build" Dec 03 00:31:20 crc kubenswrapper[4953]: I1203 00:31:20.011191 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/2b0995a0-793c-4aac-9d8d-bc0dd4604b29-buildcachedir\") pod \"sg-bridge-2-build\" (UID: \"2b0995a0-793c-4aac-9d8d-bc0dd4604b29\") " pod="service-telemetry/sg-bridge-2-build" Dec 03 00:31:20 crc kubenswrapper[4953]: I1203 00:31:20.011218 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/2b0995a0-793c-4aac-9d8d-bc0dd4604b29-build-ca-bundles\") pod \"sg-bridge-2-build\" (UID: \"2b0995a0-793c-4aac-9d8d-bc0dd4604b29\") " pod="service-telemetry/sg-bridge-2-build" Dec 03 00:31:20 crc kubenswrapper[4953]: I1203 00:31:20.011277 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"builder-dockercfg-9prwd-push\" (UniqueName: \"kubernetes.io/secret/2b0995a0-793c-4aac-9d8d-bc0dd4604b29-builder-dockercfg-9prwd-push\") pod \"sg-bridge-2-build\" (UID: \"2b0995a0-793c-4aac-9d8d-bc0dd4604b29\") " pod="service-telemetry/sg-bridge-2-build" Dec 03 00:31:20 crc kubenswrapper[4953]: I1203 00:31:20.011300 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/2b0995a0-793c-4aac-9d8d-bc0dd4604b29-build-proxy-ca-bundles\") pod \"sg-bridge-2-build\" (UID: \"2b0995a0-793c-4aac-9d8d-bc0dd4604b29\") " pod="service-telemetry/sg-bridge-2-build" Dec 03 00:31:20 crc kubenswrapper[4953]: I1203 00:31:20.011327 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/2b0995a0-793c-4aac-9d8d-bc0dd4604b29-container-storage-run\") pod \"sg-bridge-2-build\" (UID: \"2b0995a0-793c-4aac-9d8d-bc0dd4604b29\") " pod="service-telemetry/sg-bridge-2-build" Dec 03 00:31:20 crc kubenswrapper[4953]: I1203 00:31:20.011838 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/2b0995a0-793c-4aac-9d8d-bc0dd4604b29-buildworkdir\") pod \"sg-bridge-2-build\" (UID: \"2b0995a0-793c-4aac-9d8d-bc0dd4604b29\") " pod="service-telemetry/sg-bridge-2-build" Dec 03 00:31:20 crc kubenswrapper[4953]: I1203 00:31:20.011925 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/2b0995a0-793c-4aac-9d8d-bc0dd4604b29-node-pullsecrets\") pod \"sg-bridge-2-build\" (UID: \"2b0995a0-793c-4aac-9d8d-bc0dd4604b29\") " pod="service-telemetry/sg-bridge-2-build" Dec 03 00:31:20 crc kubenswrapper[4953]: I1203 00:31:20.012051 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/2b0995a0-793c-4aac-9d8d-bc0dd4604b29-build-blob-cache\") pod \"sg-bridge-2-build\" (UID: \"2b0995a0-793c-4aac-9d8d-bc0dd4604b29\") " pod="service-telemetry/sg-bridge-2-build" Dec 03 00:31:20 crc kubenswrapper[4953]: I1203 00:31:20.011882 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/2b0995a0-793c-4aac-9d8d-bc0dd4604b29-container-storage-run\") pod \"sg-bridge-2-build\" (UID: \"2b0995a0-793c-4aac-9d8d-bc0dd4604b29\") " pod="service-telemetry/sg-bridge-2-build" Dec 03 00:31:20 crc kubenswrapper[4953]: I1203 00:31:20.012484 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/2b0995a0-793c-4aac-9d8d-bc0dd4604b29-container-storage-root\") pod \"sg-bridge-2-build\" (UID: \"2b0995a0-793c-4aac-9d8d-bc0dd4604b29\") " pod="service-telemetry/sg-bridge-2-build" Dec 03 00:31:20 crc kubenswrapper[4953]: I1203 00:31:20.012888 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/2b0995a0-793c-4aac-9d8d-bc0dd4604b29-build-system-configs\") pod \"sg-bridge-2-build\" (UID: \"2b0995a0-793c-4aac-9d8d-bc0dd4604b29\") " pod="service-telemetry/sg-bridge-2-build" Dec 03 00:31:20 crc kubenswrapper[4953]: I1203 00:31:20.013083 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/2b0995a0-793c-4aac-9d8d-bc0dd4604b29-buildcachedir\") pod \"sg-bridge-2-build\" (UID: \"2b0995a0-793c-4aac-9d8d-bc0dd4604b29\") " pod="service-telemetry/sg-bridge-2-build" Dec 03 00:31:20 crc kubenswrapper[4953]: I1203 00:31:20.013210 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/2b0995a0-793c-4aac-9d8d-bc0dd4604b29-build-ca-bundles\") pod \"sg-bridge-2-build\" (UID: \"2b0995a0-793c-4aac-9d8d-bc0dd4604b29\") " pod="service-telemetry/sg-bridge-2-build" Dec 03 00:31:20 crc kubenswrapper[4953]: I1203 00:31:20.013231 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/2b0995a0-793c-4aac-9d8d-bc0dd4604b29-build-proxy-ca-bundles\") pod \"sg-bridge-2-build\" (UID: \"2b0995a0-793c-4aac-9d8d-bc0dd4604b29\") " pod="service-telemetry/sg-bridge-2-build" Dec 03 00:31:20 crc kubenswrapper[4953]: I1203 00:31:20.020820 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"builder-dockercfg-9prwd-pull\" (UniqueName: \"kubernetes.io/secret/2b0995a0-793c-4aac-9d8d-bc0dd4604b29-builder-dockercfg-9prwd-pull\") pod \"sg-bridge-2-build\" (UID: \"2b0995a0-793c-4aac-9d8d-bc0dd4604b29\") " pod="service-telemetry/sg-bridge-2-build" Dec 03 00:31:20 crc kubenswrapper[4953]: I1203 00:31:20.020893 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"builder-dockercfg-9prwd-push\" (UniqueName: \"kubernetes.io/secret/2b0995a0-793c-4aac-9d8d-bc0dd4604b29-builder-dockercfg-9prwd-push\") pod \"sg-bridge-2-build\" (UID: \"2b0995a0-793c-4aac-9d8d-bc0dd4604b29\") " pod="service-telemetry/sg-bridge-2-build" Dec 03 00:31:20 crc kubenswrapper[4953]: I1203 00:31:20.039942 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bxf8m\" (UniqueName: \"kubernetes.io/projected/2b0995a0-793c-4aac-9d8d-bc0dd4604b29-kube-api-access-bxf8m\") pod \"sg-bridge-2-build\" (UID: \"2b0995a0-793c-4aac-9d8d-bc0dd4604b29\") " pod="service-telemetry/sg-bridge-2-build" Dec 03 00:31:20 crc kubenswrapper[4953]: I1203 00:31:20.132800 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/sg-bridge-2-build" Dec 03 00:31:20 crc kubenswrapper[4953]: I1203 00:31:20.374555 4953 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/sg-bridge-2-build"] Dec 03 00:31:21 crc kubenswrapper[4953]: I1203 00:31:21.034451 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/sg-bridge-2-build" event={"ID":"2b0995a0-793c-4aac-9d8d-bc0dd4604b29","Type":"ContainerStarted","Data":"f5b246b06ab6acf807983692c59d925a4749488c2261a73a88711bc2a9a122bc"} Dec 03 00:31:21 crc kubenswrapper[4953]: I1203 00:31:21.034924 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/sg-bridge-2-build" event={"ID":"2b0995a0-793c-4aac-9d8d-bc0dd4604b29","Type":"ContainerStarted","Data":"88a09ce6d3306bc31625f757e0719150cf141e792d572a2571384dae68438517"} Dec 03 00:31:22 crc kubenswrapper[4953]: I1203 00:31:22.045167 4953 generic.go:334] "Generic (PLEG): container finished" podID="2b0995a0-793c-4aac-9d8d-bc0dd4604b29" containerID="f5b246b06ab6acf807983692c59d925a4749488c2261a73a88711bc2a9a122bc" exitCode=0 Dec 03 00:31:22 crc kubenswrapper[4953]: I1203 00:31:22.045290 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/sg-bridge-2-build" event={"ID":"2b0995a0-793c-4aac-9d8d-bc0dd4604b29","Type":"ContainerDied","Data":"f5b246b06ab6acf807983692c59d925a4749488c2261a73a88711bc2a9a122bc"} Dec 03 00:31:23 crc kubenswrapper[4953]: I1203 00:31:23.056762 4953 generic.go:334] "Generic (PLEG): container finished" podID="2b0995a0-793c-4aac-9d8d-bc0dd4604b29" containerID="465e86e3fab97f2e5b2ec686df89e6b03e90bcb4333712ecf58b5894e55d88fa" exitCode=0 Dec 03 00:31:23 crc kubenswrapper[4953]: I1203 00:31:23.057242 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/sg-bridge-2-build" event={"ID":"2b0995a0-793c-4aac-9d8d-bc0dd4604b29","Type":"ContainerDied","Data":"465e86e3fab97f2e5b2ec686df89e6b03e90bcb4333712ecf58b5894e55d88fa"} Dec 03 00:31:23 crc kubenswrapper[4953]: I1203 00:31:23.107763 4953 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_sg-bridge-2-build_2b0995a0-793c-4aac-9d8d-bc0dd4604b29/manage-dockerfile/0.log" Dec 03 00:31:24 crc kubenswrapper[4953]: I1203 00:31:24.069732 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/sg-bridge-2-build" event={"ID":"2b0995a0-793c-4aac-9d8d-bc0dd4604b29","Type":"ContainerStarted","Data":"d87e2d728d2db4ad2c8cca90914c44175cd07db42b324bdcba1a2ae1e25cb573"} Dec 03 00:31:24 crc kubenswrapper[4953]: I1203 00:31:24.098448 4953 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="service-telemetry/sg-bridge-2-build" podStartSLOduration=5.098421874 podStartE2EDuration="5.098421874s" podCreationTimestamp="2025-12-03 00:31:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 00:31:24.096292102 +0000 UTC m=+1473.080572925" watchObservedRunningTime="2025-12-03 00:31:24.098421874 +0000 UTC m=+1473.082702677" Dec 03 00:31:26 crc kubenswrapper[4953]: I1203 00:31:26.408428 4953 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-7t8d6"] Dec 03 00:31:26 crc kubenswrapper[4953]: I1203 00:31:26.410342 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-7t8d6" Dec 03 00:31:26 crc kubenswrapper[4953]: I1203 00:31:26.428068 4953 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-7t8d6"] Dec 03 00:31:26 crc kubenswrapper[4953]: I1203 00:31:26.527648 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a398a67b-f052-465e-8b7f-3913a21f1adf-utilities\") pod \"redhat-operators-7t8d6\" (UID: \"a398a67b-f052-465e-8b7f-3913a21f1adf\") " pod="openshift-marketplace/redhat-operators-7t8d6" Dec 03 00:31:26 crc kubenswrapper[4953]: I1203 00:31:26.528332 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a398a67b-f052-465e-8b7f-3913a21f1adf-catalog-content\") pod \"redhat-operators-7t8d6\" (UID: \"a398a67b-f052-465e-8b7f-3913a21f1adf\") " pod="openshift-marketplace/redhat-operators-7t8d6" Dec 03 00:31:26 crc kubenswrapper[4953]: I1203 00:31:26.528511 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n84wb\" (UniqueName: \"kubernetes.io/projected/a398a67b-f052-465e-8b7f-3913a21f1adf-kube-api-access-n84wb\") pod \"redhat-operators-7t8d6\" (UID: \"a398a67b-f052-465e-8b7f-3913a21f1adf\") " pod="openshift-marketplace/redhat-operators-7t8d6" Dec 03 00:31:26 crc kubenswrapper[4953]: I1203 00:31:26.630650 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a398a67b-f052-465e-8b7f-3913a21f1adf-catalog-content\") pod \"redhat-operators-7t8d6\" (UID: \"a398a67b-f052-465e-8b7f-3913a21f1adf\") " pod="openshift-marketplace/redhat-operators-7t8d6" Dec 03 00:31:26 crc kubenswrapper[4953]: I1203 00:31:26.630710 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n84wb\" (UniqueName: \"kubernetes.io/projected/a398a67b-f052-465e-8b7f-3913a21f1adf-kube-api-access-n84wb\") pod \"redhat-operators-7t8d6\" (UID: \"a398a67b-f052-465e-8b7f-3913a21f1adf\") " pod="openshift-marketplace/redhat-operators-7t8d6" Dec 03 00:31:26 crc kubenswrapper[4953]: I1203 00:31:26.630780 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a398a67b-f052-465e-8b7f-3913a21f1adf-utilities\") pod \"redhat-operators-7t8d6\" (UID: \"a398a67b-f052-465e-8b7f-3913a21f1adf\") " pod="openshift-marketplace/redhat-operators-7t8d6" Dec 03 00:31:26 crc kubenswrapper[4953]: I1203 00:31:26.631564 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a398a67b-f052-465e-8b7f-3913a21f1adf-catalog-content\") pod \"redhat-operators-7t8d6\" (UID: \"a398a67b-f052-465e-8b7f-3913a21f1adf\") " pod="openshift-marketplace/redhat-operators-7t8d6" Dec 03 00:31:26 crc kubenswrapper[4953]: I1203 00:31:26.631724 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a398a67b-f052-465e-8b7f-3913a21f1adf-utilities\") pod \"redhat-operators-7t8d6\" (UID: \"a398a67b-f052-465e-8b7f-3913a21f1adf\") " pod="openshift-marketplace/redhat-operators-7t8d6" Dec 03 00:31:26 crc kubenswrapper[4953]: I1203 00:31:26.655747 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n84wb\" (UniqueName: \"kubernetes.io/projected/a398a67b-f052-465e-8b7f-3913a21f1adf-kube-api-access-n84wb\") pod \"redhat-operators-7t8d6\" (UID: \"a398a67b-f052-465e-8b7f-3913a21f1adf\") " pod="openshift-marketplace/redhat-operators-7t8d6" Dec 03 00:31:26 crc kubenswrapper[4953]: I1203 00:31:26.728011 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-7t8d6" Dec 03 00:31:27 crc kubenswrapper[4953]: I1203 00:31:27.019063 4953 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-7t8d6"] Dec 03 00:31:27 crc kubenswrapper[4953]: I1203 00:31:27.092005 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7t8d6" event={"ID":"a398a67b-f052-465e-8b7f-3913a21f1adf","Type":"ContainerStarted","Data":"96eaffa2cc1edcc723b44143a9c30e0118d2a986cd68a044af226736c8952985"} Dec 03 00:31:28 crc kubenswrapper[4953]: I1203 00:31:28.101621 4953 generic.go:334] "Generic (PLEG): container finished" podID="a398a67b-f052-465e-8b7f-3913a21f1adf" containerID="be0cf69ced4774b5ef586382607eb3411c04472dc9beca8ee7a540ca915ce919" exitCode=0 Dec 03 00:31:28 crc kubenswrapper[4953]: I1203 00:31:28.101678 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7t8d6" event={"ID":"a398a67b-f052-465e-8b7f-3913a21f1adf","Type":"ContainerDied","Data":"be0cf69ced4774b5ef586382607eb3411c04472dc9beca8ee7a540ca915ce919"} Dec 03 00:31:28 crc kubenswrapper[4953]: I1203 00:31:28.103906 4953 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 03 00:31:29 crc kubenswrapper[4953]: I1203 00:31:29.112176 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7t8d6" event={"ID":"a398a67b-f052-465e-8b7f-3913a21f1adf","Type":"ContainerStarted","Data":"9390e28cb51ce6a9f902aec8b2a20205d250ebad4ab2c1f81b21abbb67083109"} Dec 03 00:31:31 crc kubenswrapper[4953]: I1203 00:31:31.131013 4953 generic.go:334] "Generic (PLEG): container finished" podID="a398a67b-f052-465e-8b7f-3913a21f1adf" containerID="9390e28cb51ce6a9f902aec8b2a20205d250ebad4ab2c1f81b21abbb67083109" exitCode=0 Dec 03 00:31:31 crc kubenswrapper[4953]: I1203 00:31:31.131097 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7t8d6" event={"ID":"a398a67b-f052-465e-8b7f-3913a21f1adf","Type":"ContainerDied","Data":"9390e28cb51ce6a9f902aec8b2a20205d250ebad4ab2c1f81b21abbb67083109"} Dec 03 00:31:32 crc kubenswrapper[4953]: I1203 00:31:32.146179 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7t8d6" event={"ID":"a398a67b-f052-465e-8b7f-3913a21f1adf","Type":"ContainerStarted","Data":"8fa4ad257f34f8d780a0ac2b0e1e09f742584890fa8e105dba165e4629138e1c"} Dec 03 00:31:32 crc kubenswrapper[4953]: I1203 00:31:32.174189 4953 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-7t8d6" podStartSLOduration=2.7015673639999997 podStartE2EDuration="6.174152735s" podCreationTimestamp="2025-12-03 00:31:26 +0000 UTC" firstStartedPulling="2025-12-03 00:31:28.103596577 +0000 UTC m=+1477.087877360" lastFinishedPulling="2025-12-03 00:31:31.576181948 +0000 UTC m=+1480.560462731" observedRunningTime="2025-12-03 00:31:32.168349832 +0000 UTC m=+1481.152630615" watchObservedRunningTime="2025-12-03 00:31:32.174152735 +0000 UTC m=+1481.158433518" Dec 03 00:31:35 crc kubenswrapper[4953]: I1203 00:31:35.656349 4953 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-6dh7f"] Dec 03 00:31:35 crc kubenswrapper[4953]: I1203 00:31:35.658069 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-6dh7f" Dec 03 00:31:35 crc kubenswrapper[4953]: I1203 00:31:35.670897 4953 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-6dh7f"] Dec 03 00:31:35 crc kubenswrapper[4953]: I1203 00:31:35.807011 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jlbsb\" (UniqueName: \"kubernetes.io/projected/c9f7486f-c563-434d-b202-9211da3efe5c-kube-api-access-jlbsb\") pod \"certified-operators-6dh7f\" (UID: \"c9f7486f-c563-434d-b202-9211da3efe5c\") " pod="openshift-marketplace/certified-operators-6dh7f" Dec 03 00:31:35 crc kubenswrapper[4953]: I1203 00:31:35.807090 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c9f7486f-c563-434d-b202-9211da3efe5c-catalog-content\") pod \"certified-operators-6dh7f\" (UID: \"c9f7486f-c563-434d-b202-9211da3efe5c\") " pod="openshift-marketplace/certified-operators-6dh7f" Dec 03 00:31:35 crc kubenswrapper[4953]: I1203 00:31:35.807295 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c9f7486f-c563-434d-b202-9211da3efe5c-utilities\") pod \"certified-operators-6dh7f\" (UID: \"c9f7486f-c563-434d-b202-9211da3efe5c\") " pod="openshift-marketplace/certified-operators-6dh7f" Dec 03 00:31:35 crc kubenswrapper[4953]: I1203 00:31:35.909277 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c9f7486f-c563-434d-b202-9211da3efe5c-catalog-content\") pod \"certified-operators-6dh7f\" (UID: \"c9f7486f-c563-434d-b202-9211da3efe5c\") " pod="openshift-marketplace/certified-operators-6dh7f" Dec 03 00:31:35 crc kubenswrapper[4953]: I1203 00:31:35.909368 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c9f7486f-c563-434d-b202-9211da3efe5c-utilities\") pod \"certified-operators-6dh7f\" (UID: \"c9f7486f-c563-434d-b202-9211da3efe5c\") " pod="openshift-marketplace/certified-operators-6dh7f" Dec 03 00:31:35 crc kubenswrapper[4953]: I1203 00:31:35.909419 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jlbsb\" (UniqueName: \"kubernetes.io/projected/c9f7486f-c563-434d-b202-9211da3efe5c-kube-api-access-jlbsb\") pod \"certified-operators-6dh7f\" (UID: \"c9f7486f-c563-434d-b202-9211da3efe5c\") " pod="openshift-marketplace/certified-operators-6dh7f" Dec 03 00:31:35 crc kubenswrapper[4953]: I1203 00:31:35.909993 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c9f7486f-c563-434d-b202-9211da3efe5c-catalog-content\") pod \"certified-operators-6dh7f\" (UID: \"c9f7486f-c563-434d-b202-9211da3efe5c\") " pod="openshift-marketplace/certified-operators-6dh7f" Dec 03 00:31:35 crc kubenswrapper[4953]: I1203 00:31:35.910181 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c9f7486f-c563-434d-b202-9211da3efe5c-utilities\") pod \"certified-operators-6dh7f\" (UID: \"c9f7486f-c563-434d-b202-9211da3efe5c\") " pod="openshift-marketplace/certified-operators-6dh7f" Dec 03 00:31:35 crc kubenswrapper[4953]: I1203 00:31:35.935134 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jlbsb\" (UniqueName: \"kubernetes.io/projected/c9f7486f-c563-434d-b202-9211da3efe5c-kube-api-access-jlbsb\") pod \"certified-operators-6dh7f\" (UID: \"c9f7486f-c563-434d-b202-9211da3efe5c\") " pod="openshift-marketplace/certified-operators-6dh7f" Dec 03 00:31:36 crc kubenswrapper[4953]: I1203 00:31:36.014530 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-6dh7f" Dec 03 00:31:36 crc kubenswrapper[4953]: I1203 00:31:36.350844 4953 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-6dh7f"] Dec 03 00:31:36 crc kubenswrapper[4953]: W1203 00:31:36.366347 4953 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc9f7486f_c563_434d_b202_9211da3efe5c.slice/crio-32e58c65f92695433348820911e06b7c17e4f2ec6f4fe80ebe0d60815ee8b8eb WatchSource:0}: Error finding container 32e58c65f92695433348820911e06b7c17e4f2ec6f4fe80ebe0d60815ee8b8eb: Status 404 returned error can't find the container with id 32e58c65f92695433348820911e06b7c17e4f2ec6f4fe80ebe0d60815ee8b8eb Dec 03 00:31:36 crc kubenswrapper[4953]: I1203 00:31:36.728304 4953 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-7t8d6" Dec 03 00:31:36 crc kubenswrapper[4953]: I1203 00:31:36.728416 4953 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-7t8d6" Dec 03 00:31:37 crc kubenswrapper[4953]: I1203 00:31:37.186409 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6dh7f" event={"ID":"c9f7486f-c563-434d-b202-9211da3efe5c","Type":"ContainerStarted","Data":"32e58c65f92695433348820911e06b7c17e4f2ec6f4fe80ebe0d60815ee8b8eb"} Dec 03 00:31:37 crc kubenswrapper[4953]: I1203 00:31:37.778759 4953 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-7t8d6" podUID="a398a67b-f052-465e-8b7f-3913a21f1adf" containerName="registry-server" probeResult="failure" output=< Dec 03 00:31:37 crc kubenswrapper[4953]: timeout: failed to connect service ":50051" within 1s Dec 03 00:31:37 crc kubenswrapper[4953]: > Dec 03 00:31:38 crc kubenswrapper[4953]: I1203 00:31:38.195395 4953 generic.go:334] "Generic (PLEG): container finished" podID="c9f7486f-c563-434d-b202-9211da3efe5c" containerID="abd14b4adba74d368bdffc4e5a41576ec3b85ea8b821f4ff786590c7603965bb" exitCode=0 Dec 03 00:31:38 crc kubenswrapper[4953]: I1203 00:31:38.195447 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6dh7f" event={"ID":"c9f7486f-c563-434d-b202-9211da3efe5c","Type":"ContainerDied","Data":"abd14b4adba74d368bdffc4e5a41576ec3b85ea8b821f4ff786590c7603965bb"} Dec 03 00:31:39 crc kubenswrapper[4953]: I1203 00:31:39.210435 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6dh7f" event={"ID":"c9f7486f-c563-434d-b202-9211da3efe5c","Type":"ContainerStarted","Data":"e614e99cde4f3f7e03d0fda154a3c7c246aef60fbbc895da822370dd777bc9f0"} Dec 03 00:31:41 crc kubenswrapper[4953]: I1203 00:31:41.242632 4953 generic.go:334] "Generic (PLEG): container finished" podID="c9f7486f-c563-434d-b202-9211da3efe5c" containerID="e614e99cde4f3f7e03d0fda154a3c7c246aef60fbbc895da822370dd777bc9f0" exitCode=0 Dec 03 00:31:41 crc kubenswrapper[4953]: I1203 00:31:41.242725 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6dh7f" event={"ID":"c9f7486f-c563-434d-b202-9211da3efe5c","Type":"ContainerDied","Data":"e614e99cde4f3f7e03d0fda154a3c7c246aef60fbbc895da822370dd777bc9f0"} Dec 03 00:31:42 crc kubenswrapper[4953]: I1203 00:31:42.252899 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6dh7f" event={"ID":"c9f7486f-c563-434d-b202-9211da3efe5c","Type":"ContainerStarted","Data":"38c5244e77121ce208d245274c61fc56a3c58f0dd1ac6db60eb195436378f788"} Dec 03 00:31:42 crc kubenswrapper[4953]: I1203 00:31:42.294275 4953 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-6dh7f" podStartSLOduration=3.844583716 podStartE2EDuration="7.294251122s" podCreationTimestamp="2025-12-03 00:31:35 +0000 UTC" firstStartedPulling="2025-12-03 00:31:38.198214393 +0000 UTC m=+1487.182495176" lastFinishedPulling="2025-12-03 00:31:41.647881789 +0000 UTC m=+1490.632162582" observedRunningTime="2025-12-03 00:31:42.290545281 +0000 UTC m=+1491.274826094" watchObservedRunningTime="2025-12-03 00:31:42.294251122 +0000 UTC m=+1491.278531905" Dec 03 00:31:46 crc kubenswrapper[4953]: I1203 00:31:46.015626 4953 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-6dh7f" Dec 03 00:31:46 crc kubenswrapper[4953]: I1203 00:31:46.016253 4953 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-6dh7f" Dec 03 00:31:46 crc kubenswrapper[4953]: I1203 00:31:46.069722 4953 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-6dh7f" Dec 03 00:31:46 crc kubenswrapper[4953]: I1203 00:31:46.347066 4953 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-6dh7f" Dec 03 00:31:46 crc kubenswrapper[4953]: I1203 00:31:46.775934 4953 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-7t8d6" Dec 03 00:31:46 crc kubenswrapper[4953]: I1203 00:31:46.827676 4953 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-7t8d6" Dec 03 00:31:49 crc kubenswrapper[4953]: I1203 00:31:49.634469 4953 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-6dh7f"] Dec 03 00:31:49 crc kubenswrapper[4953]: I1203 00:31:49.634732 4953 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-6dh7f" podUID="c9f7486f-c563-434d-b202-9211da3efe5c" containerName="registry-server" containerID="cri-o://38c5244e77121ce208d245274c61fc56a3c58f0dd1ac6db60eb195436378f788" gracePeriod=2 Dec 03 00:31:50 crc kubenswrapper[4953]: I1203 00:31:50.235791 4953 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-7t8d6"] Dec 03 00:31:50 crc kubenswrapper[4953]: I1203 00:31:50.236778 4953 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-7t8d6" podUID="a398a67b-f052-465e-8b7f-3913a21f1adf" containerName="registry-server" containerID="cri-o://8fa4ad257f34f8d780a0ac2b0e1e09f742584890fa8e105dba165e4629138e1c" gracePeriod=2 Dec 03 00:31:51 crc kubenswrapper[4953]: I1203 00:31:51.159890 4953 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-6dh7f" Dec 03 00:31:51 crc kubenswrapper[4953]: I1203 00:31:51.270902 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jlbsb\" (UniqueName: \"kubernetes.io/projected/c9f7486f-c563-434d-b202-9211da3efe5c-kube-api-access-jlbsb\") pod \"c9f7486f-c563-434d-b202-9211da3efe5c\" (UID: \"c9f7486f-c563-434d-b202-9211da3efe5c\") " Dec 03 00:31:51 crc kubenswrapper[4953]: I1203 00:31:51.271081 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c9f7486f-c563-434d-b202-9211da3efe5c-utilities\") pod \"c9f7486f-c563-434d-b202-9211da3efe5c\" (UID: \"c9f7486f-c563-434d-b202-9211da3efe5c\") " Dec 03 00:31:51 crc kubenswrapper[4953]: I1203 00:31:51.271150 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c9f7486f-c563-434d-b202-9211da3efe5c-catalog-content\") pod \"c9f7486f-c563-434d-b202-9211da3efe5c\" (UID: \"c9f7486f-c563-434d-b202-9211da3efe5c\") " Dec 03 00:31:51 crc kubenswrapper[4953]: I1203 00:31:51.272023 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c9f7486f-c563-434d-b202-9211da3efe5c-utilities" (OuterVolumeSpecName: "utilities") pod "c9f7486f-c563-434d-b202-9211da3efe5c" (UID: "c9f7486f-c563-434d-b202-9211da3efe5c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 00:31:51 crc kubenswrapper[4953]: I1203 00:31:51.278473 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c9f7486f-c563-434d-b202-9211da3efe5c-kube-api-access-jlbsb" (OuterVolumeSpecName: "kube-api-access-jlbsb") pod "c9f7486f-c563-434d-b202-9211da3efe5c" (UID: "c9f7486f-c563-434d-b202-9211da3efe5c"). InnerVolumeSpecName "kube-api-access-jlbsb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:31:51 crc kubenswrapper[4953]: I1203 00:31:51.309722 4953 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-7t8d6" Dec 03 00:31:51 crc kubenswrapper[4953]: I1203 00:31:51.317315 4953 generic.go:334] "Generic (PLEG): container finished" podID="c9f7486f-c563-434d-b202-9211da3efe5c" containerID="38c5244e77121ce208d245274c61fc56a3c58f0dd1ac6db60eb195436378f788" exitCode=0 Dec 03 00:31:51 crc kubenswrapper[4953]: I1203 00:31:51.317378 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6dh7f" event={"ID":"c9f7486f-c563-434d-b202-9211da3efe5c","Type":"ContainerDied","Data":"38c5244e77121ce208d245274c61fc56a3c58f0dd1ac6db60eb195436378f788"} Dec 03 00:31:51 crc kubenswrapper[4953]: I1203 00:31:51.317455 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6dh7f" event={"ID":"c9f7486f-c563-434d-b202-9211da3efe5c","Type":"ContainerDied","Data":"32e58c65f92695433348820911e06b7c17e4f2ec6f4fe80ebe0d60815ee8b8eb"} Dec 03 00:31:51 crc kubenswrapper[4953]: I1203 00:31:51.317489 4953 scope.go:117] "RemoveContainer" containerID="38c5244e77121ce208d245274c61fc56a3c58f0dd1ac6db60eb195436378f788" Dec 03 00:31:51 crc kubenswrapper[4953]: I1203 00:31:51.317496 4953 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-6dh7f" Dec 03 00:31:51 crc kubenswrapper[4953]: I1203 00:31:51.320653 4953 generic.go:334] "Generic (PLEG): container finished" podID="a398a67b-f052-465e-8b7f-3913a21f1adf" containerID="8fa4ad257f34f8d780a0ac2b0e1e09f742584890fa8e105dba165e4629138e1c" exitCode=0 Dec 03 00:31:51 crc kubenswrapper[4953]: I1203 00:31:51.320697 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7t8d6" event={"ID":"a398a67b-f052-465e-8b7f-3913a21f1adf","Type":"ContainerDied","Data":"8fa4ad257f34f8d780a0ac2b0e1e09f742584890fa8e105dba165e4629138e1c"} Dec 03 00:31:51 crc kubenswrapper[4953]: I1203 00:31:51.320719 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7t8d6" event={"ID":"a398a67b-f052-465e-8b7f-3913a21f1adf","Type":"ContainerDied","Data":"96eaffa2cc1edcc723b44143a9c30e0118d2a986cd68a044af226736c8952985"} Dec 03 00:31:51 crc kubenswrapper[4953]: I1203 00:31:51.320836 4953 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-7t8d6" Dec 03 00:31:51 crc kubenswrapper[4953]: I1203 00:31:51.333105 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c9f7486f-c563-434d-b202-9211da3efe5c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c9f7486f-c563-434d-b202-9211da3efe5c" (UID: "c9f7486f-c563-434d-b202-9211da3efe5c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 00:31:51 crc kubenswrapper[4953]: I1203 00:31:51.351343 4953 scope.go:117] "RemoveContainer" containerID="e614e99cde4f3f7e03d0fda154a3c7c246aef60fbbc895da822370dd777bc9f0" Dec 03 00:31:51 crc kubenswrapper[4953]: I1203 00:31:51.373315 4953 scope.go:117] "RemoveContainer" containerID="abd14b4adba74d368bdffc4e5a41576ec3b85ea8b821f4ff786590c7603965bb" Dec 03 00:31:51 crc kubenswrapper[4953]: I1203 00:31:51.375026 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a398a67b-f052-465e-8b7f-3913a21f1adf-utilities\") pod \"a398a67b-f052-465e-8b7f-3913a21f1adf\" (UID: \"a398a67b-f052-465e-8b7f-3913a21f1adf\") " Dec 03 00:31:51 crc kubenswrapper[4953]: I1203 00:31:51.375160 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a398a67b-f052-465e-8b7f-3913a21f1adf-catalog-content\") pod \"a398a67b-f052-465e-8b7f-3913a21f1adf\" (UID: \"a398a67b-f052-465e-8b7f-3913a21f1adf\") " Dec 03 00:31:51 crc kubenswrapper[4953]: I1203 00:31:51.375211 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n84wb\" (UniqueName: \"kubernetes.io/projected/a398a67b-f052-465e-8b7f-3913a21f1adf-kube-api-access-n84wb\") pod \"a398a67b-f052-465e-8b7f-3913a21f1adf\" (UID: \"a398a67b-f052-465e-8b7f-3913a21f1adf\") " Dec 03 00:31:51 crc kubenswrapper[4953]: I1203 00:31:51.377466 4953 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jlbsb\" (UniqueName: \"kubernetes.io/projected/c9f7486f-c563-434d-b202-9211da3efe5c-kube-api-access-jlbsb\") on node \"crc\" DevicePath \"\"" Dec 03 00:31:51 crc kubenswrapper[4953]: I1203 00:31:51.377498 4953 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c9f7486f-c563-434d-b202-9211da3efe5c-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 00:31:51 crc kubenswrapper[4953]: I1203 00:31:51.377512 4953 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c9f7486f-c563-434d-b202-9211da3efe5c-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 00:31:51 crc kubenswrapper[4953]: I1203 00:31:51.384131 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a398a67b-f052-465e-8b7f-3913a21f1adf-utilities" (OuterVolumeSpecName: "utilities") pod "a398a67b-f052-465e-8b7f-3913a21f1adf" (UID: "a398a67b-f052-465e-8b7f-3913a21f1adf"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 00:31:51 crc kubenswrapper[4953]: I1203 00:31:51.390986 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a398a67b-f052-465e-8b7f-3913a21f1adf-kube-api-access-n84wb" (OuterVolumeSpecName: "kube-api-access-n84wb") pod "a398a67b-f052-465e-8b7f-3913a21f1adf" (UID: "a398a67b-f052-465e-8b7f-3913a21f1adf"). InnerVolumeSpecName "kube-api-access-n84wb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:31:51 crc kubenswrapper[4953]: I1203 00:31:51.402130 4953 scope.go:117] "RemoveContainer" containerID="38c5244e77121ce208d245274c61fc56a3c58f0dd1ac6db60eb195436378f788" Dec 03 00:31:51 crc kubenswrapper[4953]: E1203 00:31:51.403002 4953 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"38c5244e77121ce208d245274c61fc56a3c58f0dd1ac6db60eb195436378f788\": container with ID starting with 38c5244e77121ce208d245274c61fc56a3c58f0dd1ac6db60eb195436378f788 not found: ID does not exist" containerID="38c5244e77121ce208d245274c61fc56a3c58f0dd1ac6db60eb195436378f788" Dec 03 00:31:51 crc kubenswrapper[4953]: I1203 00:31:51.403116 4953 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"38c5244e77121ce208d245274c61fc56a3c58f0dd1ac6db60eb195436378f788"} err="failed to get container status \"38c5244e77121ce208d245274c61fc56a3c58f0dd1ac6db60eb195436378f788\": rpc error: code = NotFound desc = could not find container \"38c5244e77121ce208d245274c61fc56a3c58f0dd1ac6db60eb195436378f788\": container with ID starting with 38c5244e77121ce208d245274c61fc56a3c58f0dd1ac6db60eb195436378f788 not found: ID does not exist" Dec 03 00:31:51 crc kubenswrapper[4953]: I1203 00:31:51.403170 4953 scope.go:117] "RemoveContainer" containerID="e614e99cde4f3f7e03d0fda154a3c7c246aef60fbbc895da822370dd777bc9f0" Dec 03 00:31:51 crc kubenswrapper[4953]: E1203 00:31:51.403442 4953 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e614e99cde4f3f7e03d0fda154a3c7c246aef60fbbc895da822370dd777bc9f0\": container with ID starting with e614e99cde4f3f7e03d0fda154a3c7c246aef60fbbc895da822370dd777bc9f0 not found: ID does not exist" containerID="e614e99cde4f3f7e03d0fda154a3c7c246aef60fbbc895da822370dd777bc9f0" Dec 03 00:31:51 crc kubenswrapper[4953]: I1203 00:31:51.403473 4953 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e614e99cde4f3f7e03d0fda154a3c7c246aef60fbbc895da822370dd777bc9f0"} err="failed to get container status \"e614e99cde4f3f7e03d0fda154a3c7c246aef60fbbc895da822370dd777bc9f0\": rpc error: code = NotFound desc = could not find container \"e614e99cde4f3f7e03d0fda154a3c7c246aef60fbbc895da822370dd777bc9f0\": container with ID starting with e614e99cde4f3f7e03d0fda154a3c7c246aef60fbbc895da822370dd777bc9f0 not found: ID does not exist" Dec 03 00:31:51 crc kubenswrapper[4953]: I1203 00:31:51.403494 4953 scope.go:117] "RemoveContainer" containerID="abd14b4adba74d368bdffc4e5a41576ec3b85ea8b821f4ff786590c7603965bb" Dec 03 00:31:51 crc kubenswrapper[4953]: E1203 00:31:51.404651 4953 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"abd14b4adba74d368bdffc4e5a41576ec3b85ea8b821f4ff786590c7603965bb\": container with ID starting with abd14b4adba74d368bdffc4e5a41576ec3b85ea8b821f4ff786590c7603965bb not found: ID does not exist" containerID="abd14b4adba74d368bdffc4e5a41576ec3b85ea8b821f4ff786590c7603965bb" Dec 03 00:31:51 crc kubenswrapper[4953]: I1203 00:31:51.404675 4953 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"abd14b4adba74d368bdffc4e5a41576ec3b85ea8b821f4ff786590c7603965bb"} err="failed to get container status \"abd14b4adba74d368bdffc4e5a41576ec3b85ea8b821f4ff786590c7603965bb\": rpc error: code = NotFound desc = could not find container \"abd14b4adba74d368bdffc4e5a41576ec3b85ea8b821f4ff786590c7603965bb\": container with ID starting with abd14b4adba74d368bdffc4e5a41576ec3b85ea8b821f4ff786590c7603965bb not found: ID does not exist" Dec 03 00:31:51 crc kubenswrapper[4953]: I1203 00:31:51.404689 4953 scope.go:117] "RemoveContainer" containerID="8fa4ad257f34f8d780a0ac2b0e1e09f742584890fa8e105dba165e4629138e1c" Dec 03 00:31:51 crc kubenswrapper[4953]: I1203 00:31:51.423391 4953 scope.go:117] "RemoveContainer" containerID="9390e28cb51ce6a9f902aec8b2a20205d250ebad4ab2c1f81b21abbb67083109" Dec 03 00:31:51 crc kubenswrapper[4953]: I1203 00:31:51.444173 4953 scope.go:117] "RemoveContainer" containerID="be0cf69ced4774b5ef586382607eb3411c04472dc9beca8ee7a540ca915ce919" Dec 03 00:31:51 crc kubenswrapper[4953]: I1203 00:31:51.462676 4953 scope.go:117] "RemoveContainer" containerID="8fa4ad257f34f8d780a0ac2b0e1e09f742584890fa8e105dba165e4629138e1c" Dec 03 00:31:51 crc kubenswrapper[4953]: E1203 00:31:51.463372 4953 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8fa4ad257f34f8d780a0ac2b0e1e09f742584890fa8e105dba165e4629138e1c\": container with ID starting with 8fa4ad257f34f8d780a0ac2b0e1e09f742584890fa8e105dba165e4629138e1c not found: ID does not exist" containerID="8fa4ad257f34f8d780a0ac2b0e1e09f742584890fa8e105dba165e4629138e1c" Dec 03 00:31:51 crc kubenswrapper[4953]: I1203 00:31:51.463446 4953 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8fa4ad257f34f8d780a0ac2b0e1e09f742584890fa8e105dba165e4629138e1c"} err="failed to get container status \"8fa4ad257f34f8d780a0ac2b0e1e09f742584890fa8e105dba165e4629138e1c\": rpc error: code = NotFound desc = could not find container \"8fa4ad257f34f8d780a0ac2b0e1e09f742584890fa8e105dba165e4629138e1c\": container with ID starting with 8fa4ad257f34f8d780a0ac2b0e1e09f742584890fa8e105dba165e4629138e1c not found: ID does not exist" Dec 03 00:31:51 crc kubenswrapper[4953]: I1203 00:31:51.463499 4953 scope.go:117] "RemoveContainer" containerID="9390e28cb51ce6a9f902aec8b2a20205d250ebad4ab2c1f81b21abbb67083109" Dec 03 00:31:51 crc kubenswrapper[4953]: E1203 00:31:51.464197 4953 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9390e28cb51ce6a9f902aec8b2a20205d250ebad4ab2c1f81b21abbb67083109\": container with ID starting with 9390e28cb51ce6a9f902aec8b2a20205d250ebad4ab2c1f81b21abbb67083109 not found: ID does not exist" containerID="9390e28cb51ce6a9f902aec8b2a20205d250ebad4ab2c1f81b21abbb67083109" Dec 03 00:31:51 crc kubenswrapper[4953]: I1203 00:31:51.464242 4953 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9390e28cb51ce6a9f902aec8b2a20205d250ebad4ab2c1f81b21abbb67083109"} err="failed to get container status \"9390e28cb51ce6a9f902aec8b2a20205d250ebad4ab2c1f81b21abbb67083109\": rpc error: code = NotFound desc = could not find container \"9390e28cb51ce6a9f902aec8b2a20205d250ebad4ab2c1f81b21abbb67083109\": container with ID starting with 9390e28cb51ce6a9f902aec8b2a20205d250ebad4ab2c1f81b21abbb67083109 not found: ID does not exist" Dec 03 00:31:51 crc kubenswrapper[4953]: I1203 00:31:51.464279 4953 scope.go:117] "RemoveContainer" containerID="be0cf69ced4774b5ef586382607eb3411c04472dc9beca8ee7a540ca915ce919" Dec 03 00:31:51 crc kubenswrapper[4953]: E1203 00:31:51.464697 4953 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"be0cf69ced4774b5ef586382607eb3411c04472dc9beca8ee7a540ca915ce919\": container with ID starting with be0cf69ced4774b5ef586382607eb3411c04472dc9beca8ee7a540ca915ce919 not found: ID does not exist" containerID="be0cf69ced4774b5ef586382607eb3411c04472dc9beca8ee7a540ca915ce919" Dec 03 00:31:51 crc kubenswrapper[4953]: I1203 00:31:51.464780 4953 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"be0cf69ced4774b5ef586382607eb3411c04472dc9beca8ee7a540ca915ce919"} err="failed to get container status \"be0cf69ced4774b5ef586382607eb3411c04472dc9beca8ee7a540ca915ce919\": rpc error: code = NotFound desc = could not find container \"be0cf69ced4774b5ef586382607eb3411c04472dc9beca8ee7a540ca915ce919\": container with ID starting with be0cf69ced4774b5ef586382607eb3411c04472dc9beca8ee7a540ca915ce919 not found: ID does not exist" Dec 03 00:31:51 crc kubenswrapper[4953]: I1203 00:31:51.478866 4953 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a398a67b-f052-465e-8b7f-3913a21f1adf-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 00:31:51 crc kubenswrapper[4953]: I1203 00:31:51.478894 4953 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n84wb\" (UniqueName: \"kubernetes.io/projected/a398a67b-f052-465e-8b7f-3913a21f1adf-kube-api-access-n84wb\") on node \"crc\" DevicePath \"\"" Dec 03 00:31:51 crc kubenswrapper[4953]: I1203 00:31:51.512486 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a398a67b-f052-465e-8b7f-3913a21f1adf-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a398a67b-f052-465e-8b7f-3913a21f1adf" (UID: "a398a67b-f052-465e-8b7f-3913a21f1adf"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 00:31:51 crc kubenswrapper[4953]: I1203 00:31:51.580097 4953 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a398a67b-f052-465e-8b7f-3913a21f1adf-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 00:31:51 crc kubenswrapper[4953]: I1203 00:31:51.640751 4953 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-6dh7f"] Dec 03 00:31:51 crc kubenswrapper[4953]: I1203 00:31:51.645588 4953 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-6dh7f"] Dec 03 00:31:51 crc kubenswrapper[4953]: I1203 00:31:51.658784 4953 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-7t8d6"] Dec 03 00:31:51 crc kubenswrapper[4953]: I1203 00:31:51.663071 4953 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-7t8d6"] Dec 03 00:31:53 crc kubenswrapper[4953]: I1203 00:31:53.358147 4953 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a398a67b-f052-465e-8b7f-3913a21f1adf" path="/var/lib/kubelet/pods/a398a67b-f052-465e-8b7f-3913a21f1adf/volumes" Dec 03 00:31:53 crc kubenswrapper[4953]: I1203 00:31:53.358923 4953 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c9f7486f-c563-434d-b202-9211da3efe5c" path="/var/lib/kubelet/pods/c9f7486f-c563-434d-b202-9211da3efe5c/volumes" Dec 03 00:32:20 crc kubenswrapper[4953]: I1203 00:32:20.555725 4953 generic.go:334] "Generic (PLEG): container finished" podID="2b0995a0-793c-4aac-9d8d-bc0dd4604b29" containerID="d87e2d728d2db4ad2c8cca90914c44175cd07db42b324bdcba1a2ae1e25cb573" exitCode=0 Dec 03 00:32:20 crc kubenswrapper[4953]: I1203 00:32:20.555813 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/sg-bridge-2-build" event={"ID":"2b0995a0-793c-4aac-9d8d-bc0dd4604b29","Type":"ContainerDied","Data":"d87e2d728d2db4ad2c8cca90914c44175cd07db42b324bdcba1a2ae1e25cb573"} Dec 03 00:32:21 crc kubenswrapper[4953]: I1203 00:32:21.834689 4953 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/sg-bridge-2-build" Dec 03 00:32:22 crc kubenswrapper[4953]: I1203 00:32:22.022922 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"builder-dockercfg-9prwd-push\" (UniqueName: \"kubernetes.io/secret/2b0995a0-793c-4aac-9d8d-bc0dd4604b29-builder-dockercfg-9prwd-push\") pod \"2b0995a0-793c-4aac-9d8d-bc0dd4604b29\" (UID: \"2b0995a0-793c-4aac-9d8d-bc0dd4604b29\") " Dec 03 00:32:22 crc kubenswrapper[4953]: I1203 00:32:22.023017 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"builder-dockercfg-9prwd-pull\" (UniqueName: \"kubernetes.io/secret/2b0995a0-793c-4aac-9d8d-bc0dd4604b29-builder-dockercfg-9prwd-pull\") pod \"2b0995a0-793c-4aac-9d8d-bc0dd4604b29\" (UID: \"2b0995a0-793c-4aac-9d8d-bc0dd4604b29\") " Dec 03 00:32:22 crc kubenswrapper[4953]: I1203 00:32:22.023084 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/2b0995a0-793c-4aac-9d8d-bc0dd4604b29-build-system-configs\") pod \"2b0995a0-793c-4aac-9d8d-bc0dd4604b29\" (UID: \"2b0995a0-793c-4aac-9d8d-bc0dd4604b29\") " Dec 03 00:32:22 crc kubenswrapper[4953]: I1203 00:32:22.023154 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/2b0995a0-793c-4aac-9d8d-bc0dd4604b29-build-proxy-ca-bundles\") pod \"2b0995a0-793c-4aac-9d8d-bc0dd4604b29\" (UID: \"2b0995a0-793c-4aac-9d8d-bc0dd4604b29\") " Dec 03 00:32:22 crc kubenswrapper[4953]: I1203 00:32:22.023189 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/2b0995a0-793c-4aac-9d8d-bc0dd4604b29-node-pullsecrets\") pod \"2b0995a0-793c-4aac-9d8d-bc0dd4604b29\" (UID: \"2b0995a0-793c-4aac-9d8d-bc0dd4604b29\") " Dec 03 00:32:22 crc kubenswrapper[4953]: I1203 00:32:22.023218 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/2b0995a0-793c-4aac-9d8d-bc0dd4604b29-build-ca-bundles\") pod \"2b0995a0-793c-4aac-9d8d-bc0dd4604b29\" (UID: \"2b0995a0-793c-4aac-9d8d-bc0dd4604b29\") " Dec 03 00:32:22 crc kubenswrapper[4953]: I1203 00:32:22.023286 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/2b0995a0-793c-4aac-9d8d-bc0dd4604b29-container-storage-root\") pod \"2b0995a0-793c-4aac-9d8d-bc0dd4604b29\" (UID: \"2b0995a0-793c-4aac-9d8d-bc0dd4604b29\") " Dec 03 00:32:22 crc kubenswrapper[4953]: I1203 00:32:22.023328 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/2b0995a0-793c-4aac-9d8d-bc0dd4604b29-container-storage-run\") pod \"2b0995a0-793c-4aac-9d8d-bc0dd4604b29\" (UID: \"2b0995a0-793c-4aac-9d8d-bc0dd4604b29\") " Dec 03 00:32:22 crc kubenswrapper[4953]: I1203 00:32:22.023419 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/2b0995a0-793c-4aac-9d8d-bc0dd4604b29-buildworkdir\") pod \"2b0995a0-793c-4aac-9d8d-bc0dd4604b29\" (UID: \"2b0995a0-793c-4aac-9d8d-bc0dd4604b29\") " Dec 03 00:32:22 crc kubenswrapper[4953]: I1203 00:32:22.023459 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/2b0995a0-793c-4aac-9d8d-bc0dd4604b29-build-blob-cache\") pod \"2b0995a0-793c-4aac-9d8d-bc0dd4604b29\" (UID: \"2b0995a0-793c-4aac-9d8d-bc0dd4604b29\") " Dec 03 00:32:22 crc kubenswrapper[4953]: I1203 00:32:22.023514 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bxf8m\" (UniqueName: \"kubernetes.io/projected/2b0995a0-793c-4aac-9d8d-bc0dd4604b29-kube-api-access-bxf8m\") pod \"2b0995a0-793c-4aac-9d8d-bc0dd4604b29\" (UID: \"2b0995a0-793c-4aac-9d8d-bc0dd4604b29\") " Dec 03 00:32:22 crc kubenswrapper[4953]: I1203 00:32:22.023610 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/2b0995a0-793c-4aac-9d8d-bc0dd4604b29-buildcachedir\") pod \"2b0995a0-793c-4aac-9d8d-bc0dd4604b29\" (UID: \"2b0995a0-793c-4aac-9d8d-bc0dd4604b29\") " Dec 03 00:32:22 crc kubenswrapper[4953]: I1203 00:32:22.024005 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2b0995a0-793c-4aac-9d8d-bc0dd4604b29-buildcachedir" (OuterVolumeSpecName: "buildcachedir") pod "2b0995a0-793c-4aac-9d8d-bc0dd4604b29" (UID: "2b0995a0-793c-4aac-9d8d-bc0dd4604b29"). InnerVolumeSpecName "buildcachedir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 00:32:22 crc kubenswrapper[4953]: I1203 00:32:22.024657 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2b0995a0-793c-4aac-9d8d-bc0dd4604b29-build-system-configs" (OuterVolumeSpecName: "build-system-configs") pod "2b0995a0-793c-4aac-9d8d-bc0dd4604b29" (UID: "2b0995a0-793c-4aac-9d8d-bc0dd4604b29"). InnerVolumeSpecName "build-system-configs". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:32:22 crc kubenswrapper[4953]: I1203 00:32:22.026131 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2b0995a0-793c-4aac-9d8d-bc0dd4604b29-node-pullsecrets" (OuterVolumeSpecName: "node-pullsecrets") pod "2b0995a0-793c-4aac-9d8d-bc0dd4604b29" (UID: "2b0995a0-793c-4aac-9d8d-bc0dd4604b29"). InnerVolumeSpecName "node-pullsecrets". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 00:32:22 crc kubenswrapper[4953]: I1203 00:32:22.027444 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2b0995a0-793c-4aac-9d8d-bc0dd4604b29-build-ca-bundles" (OuterVolumeSpecName: "build-ca-bundles") pod "2b0995a0-793c-4aac-9d8d-bc0dd4604b29" (UID: "2b0995a0-793c-4aac-9d8d-bc0dd4604b29"). InnerVolumeSpecName "build-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:32:22 crc kubenswrapper[4953]: I1203 00:32:22.027934 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2b0995a0-793c-4aac-9d8d-bc0dd4604b29-build-proxy-ca-bundles" (OuterVolumeSpecName: "build-proxy-ca-bundles") pod "2b0995a0-793c-4aac-9d8d-bc0dd4604b29" (UID: "2b0995a0-793c-4aac-9d8d-bc0dd4604b29"). InnerVolumeSpecName "build-proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:32:22 crc kubenswrapper[4953]: I1203 00:32:22.028105 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2b0995a0-793c-4aac-9d8d-bc0dd4604b29-container-storage-run" (OuterVolumeSpecName: "container-storage-run") pod "2b0995a0-793c-4aac-9d8d-bc0dd4604b29" (UID: "2b0995a0-793c-4aac-9d8d-bc0dd4604b29"). InnerVolumeSpecName "container-storage-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 00:32:22 crc kubenswrapper[4953]: I1203 00:32:22.030350 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2b0995a0-793c-4aac-9d8d-bc0dd4604b29-buildworkdir" (OuterVolumeSpecName: "buildworkdir") pod "2b0995a0-793c-4aac-9d8d-bc0dd4604b29" (UID: "2b0995a0-793c-4aac-9d8d-bc0dd4604b29"). InnerVolumeSpecName "buildworkdir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 00:32:22 crc kubenswrapper[4953]: I1203 00:32:22.031513 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2b0995a0-793c-4aac-9d8d-bc0dd4604b29-builder-dockercfg-9prwd-push" (OuterVolumeSpecName: "builder-dockercfg-9prwd-push") pod "2b0995a0-793c-4aac-9d8d-bc0dd4604b29" (UID: "2b0995a0-793c-4aac-9d8d-bc0dd4604b29"). InnerVolumeSpecName "builder-dockercfg-9prwd-push". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:32:22 crc kubenswrapper[4953]: I1203 00:32:22.031567 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2b0995a0-793c-4aac-9d8d-bc0dd4604b29-builder-dockercfg-9prwd-pull" (OuterVolumeSpecName: "builder-dockercfg-9prwd-pull") pod "2b0995a0-793c-4aac-9d8d-bc0dd4604b29" (UID: "2b0995a0-793c-4aac-9d8d-bc0dd4604b29"). InnerVolumeSpecName "builder-dockercfg-9prwd-pull". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:32:22 crc kubenswrapper[4953]: I1203 00:32:22.034228 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2b0995a0-793c-4aac-9d8d-bc0dd4604b29-kube-api-access-bxf8m" (OuterVolumeSpecName: "kube-api-access-bxf8m") pod "2b0995a0-793c-4aac-9d8d-bc0dd4604b29" (UID: "2b0995a0-793c-4aac-9d8d-bc0dd4604b29"). InnerVolumeSpecName "kube-api-access-bxf8m". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:32:22 crc kubenswrapper[4953]: I1203 00:32:22.125932 4953 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bxf8m\" (UniqueName: \"kubernetes.io/projected/2b0995a0-793c-4aac-9d8d-bc0dd4604b29-kube-api-access-bxf8m\") on node \"crc\" DevicePath \"\"" Dec 03 00:32:22 crc kubenswrapper[4953]: I1203 00:32:22.126072 4953 reconciler_common.go:293] "Volume detached for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/2b0995a0-793c-4aac-9d8d-bc0dd4604b29-buildcachedir\") on node \"crc\" DevicePath \"\"" Dec 03 00:32:22 crc kubenswrapper[4953]: I1203 00:32:22.126098 4953 reconciler_common.go:293] "Volume detached for volume \"builder-dockercfg-9prwd-push\" (UniqueName: \"kubernetes.io/secret/2b0995a0-793c-4aac-9d8d-bc0dd4604b29-builder-dockercfg-9prwd-push\") on node \"crc\" DevicePath \"\"" Dec 03 00:32:22 crc kubenswrapper[4953]: I1203 00:32:22.126117 4953 reconciler_common.go:293] "Volume detached for volume \"builder-dockercfg-9prwd-pull\" (UniqueName: \"kubernetes.io/secret/2b0995a0-793c-4aac-9d8d-bc0dd4604b29-builder-dockercfg-9prwd-pull\") on node \"crc\" DevicePath \"\"" Dec 03 00:32:22 crc kubenswrapper[4953]: I1203 00:32:22.126137 4953 reconciler_common.go:293] "Volume detached for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/2b0995a0-793c-4aac-9d8d-bc0dd4604b29-build-system-configs\") on node \"crc\" DevicePath \"\"" Dec 03 00:32:22 crc kubenswrapper[4953]: I1203 00:32:22.126158 4953 reconciler_common.go:293] "Volume detached for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/2b0995a0-793c-4aac-9d8d-bc0dd4604b29-build-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Dec 03 00:32:22 crc kubenswrapper[4953]: I1203 00:32:22.126177 4953 reconciler_common.go:293] "Volume detached for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/2b0995a0-793c-4aac-9d8d-bc0dd4604b29-node-pullsecrets\") on node \"crc\" DevicePath \"\"" Dec 03 00:32:22 crc kubenswrapper[4953]: I1203 00:32:22.126195 4953 reconciler_common.go:293] "Volume detached for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/2b0995a0-793c-4aac-9d8d-bc0dd4604b29-build-ca-bundles\") on node \"crc\" DevicePath \"\"" Dec 03 00:32:22 crc kubenswrapper[4953]: I1203 00:32:22.126213 4953 reconciler_common.go:293] "Volume detached for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/2b0995a0-793c-4aac-9d8d-bc0dd4604b29-container-storage-run\") on node \"crc\" DevicePath \"\"" Dec 03 00:32:22 crc kubenswrapper[4953]: I1203 00:32:22.126271 4953 reconciler_common.go:293] "Volume detached for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/2b0995a0-793c-4aac-9d8d-bc0dd4604b29-buildworkdir\") on node \"crc\" DevicePath \"\"" Dec 03 00:32:22 crc kubenswrapper[4953]: I1203 00:32:22.159776 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2b0995a0-793c-4aac-9d8d-bc0dd4604b29-build-blob-cache" (OuterVolumeSpecName: "build-blob-cache") pod "2b0995a0-793c-4aac-9d8d-bc0dd4604b29" (UID: "2b0995a0-793c-4aac-9d8d-bc0dd4604b29"). InnerVolumeSpecName "build-blob-cache". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 00:32:22 crc kubenswrapper[4953]: I1203 00:32:22.227598 4953 reconciler_common.go:293] "Volume detached for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/2b0995a0-793c-4aac-9d8d-bc0dd4604b29-build-blob-cache\") on node \"crc\" DevicePath \"\"" Dec 03 00:32:22 crc kubenswrapper[4953]: I1203 00:32:22.584986 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/sg-bridge-2-build" event={"ID":"2b0995a0-793c-4aac-9d8d-bc0dd4604b29","Type":"ContainerDied","Data":"88a09ce6d3306bc31625f757e0719150cf141e792d572a2571384dae68438517"} Dec 03 00:32:22 crc kubenswrapper[4953]: I1203 00:32:22.585045 4953 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="88a09ce6d3306bc31625f757e0719150cf141e792d572a2571384dae68438517" Dec 03 00:32:22 crc kubenswrapper[4953]: I1203 00:32:22.585085 4953 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/sg-bridge-2-build" Dec 03 00:32:22 crc kubenswrapper[4953]: I1203 00:32:22.812542 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2b0995a0-793c-4aac-9d8d-bc0dd4604b29-container-storage-root" (OuterVolumeSpecName: "container-storage-root") pod "2b0995a0-793c-4aac-9d8d-bc0dd4604b29" (UID: "2b0995a0-793c-4aac-9d8d-bc0dd4604b29"). InnerVolumeSpecName "container-storage-root". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 00:32:22 crc kubenswrapper[4953]: I1203 00:32:22.840012 4953 reconciler_common.go:293] "Volume detached for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/2b0995a0-793c-4aac-9d8d-bc0dd4604b29-container-storage-root\") on node \"crc\" DevicePath \"\"" Dec 03 00:32:26 crc kubenswrapper[4953]: I1203 00:32:26.463501 4953 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["service-telemetry/prometheus-webhook-snmp-1-build"] Dec 03 00:32:26 crc kubenswrapper[4953]: E1203 00:32:26.464312 4953 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2b0995a0-793c-4aac-9d8d-bc0dd4604b29" containerName="git-clone" Dec 03 00:32:26 crc kubenswrapper[4953]: I1203 00:32:26.464329 4953 state_mem.go:107] "Deleted CPUSet assignment" podUID="2b0995a0-793c-4aac-9d8d-bc0dd4604b29" containerName="git-clone" Dec 03 00:32:26 crc kubenswrapper[4953]: E1203 00:32:26.464346 4953 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a398a67b-f052-465e-8b7f-3913a21f1adf" containerName="registry-server" Dec 03 00:32:26 crc kubenswrapper[4953]: I1203 00:32:26.464355 4953 state_mem.go:107] "Deleted CPUSet assignment" podUID="a398a67b-f052-465e-8b7f-3913a21f1adf" containerName="registry-server" Dec 03 00:32:26 crc kubenswrapper[4953]: E1203 00:32:26.464369 4953 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2b0995a0-793c-4aac-9d8d-bc0dd4604b29" containerName="manage-dockerfile" Dec 03 00:32:26 crc kubenswrapper[4953]: I1203 00:32:26.464379 4953 state_mem.go:107] "Deleted CPUSet assignment" podUID="2b0995a0-793c-4aac-9d8d-bc0dd4604b29" containerName="manage-dockerfile" Dec 03 00:32:26 crc kubenswrapper[4953]: E1203 00:32:26.464391 4953 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c9f7486f-c563-434d-b202-9211da3efe5c" containerName="extract-utilities" Dec 03 00:32:26 crc kubenswrapper[4953]: I1203 00:32:26.464398 4953 state_mem.go:107] "Deleted CPUSet assignment" podUID="c9f7486f-c563-434d-b202-9211da3efe5c" containerName="extract-utilities" Dec 03 00:32:26 crc kubenswrapper[4953]: E1203 00:32:26.464413 4953 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2b0995a0-793c-4aac-9d8d-bc0dd4604b29" containerName="docker-build" Dec 03 00:32:26 crc kubenswrapper[4953]: I1203 00:32:26.464421 4953 state_mem.go:107] "Deleted CPUSet assignment" podUID="2b0995a0-793c-4aac-9d8d-bc0dd4604b29" containerName="docker-build" Dec 03 00:32:26 crc kubenswrapper[4953]: E1203 00:32:26.464432 4953 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a398a67b-f052-465e-8b7f-3913a21f1adf" containerName="extract-content" Dec 03 00:32:26 crc kubenswrapper[4953]: I1203 00:32:26.464440 4953 state_mem.go:107] "Deleted CPUSet assignment" podUID="a398a67b-f052-465e-8b7f-3913a21f1adf" containerName="extract-content" Dec 03 00:32:26 crc kubenswrapper[4953]: E1203 00:32:26.464452 4953 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c9f7486f-c563-434d-b202-9211da3efe5c" containerName="registry-server" Dec 03 00:32:26 crc kubenswrapper[4953]: I1203 00:32:26.464460 4953 state_mem.go:107] "Deleted CPUSet assignment" podUID="c9f7486f-c563-434d-b202-9211da3efe5c" containerName="registry-server" Dec 03 00:32:26 crc kubenswrapper[4953]: E1203 00:32:26.464472 4953 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c9f7486f-c563-434d-b202-9211da3efe5c" containerName="extract-content" Dec 03 00:32:26 crc kubenswrapper[4953]: I1203 00:32:26.464480 4953 state_mem.go:107] "Deleted CPUSet assignment" podUID="c9f7486f-c563-434d-b202-9211da3efe5c" containerName="extract-content" Dec 03 00:32:26 crc kubenswrapper[4953]: E1203 00:32:26.464491 4953 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a398a67b-f052-465e-8b7f-3913a21f1adf" containerName="extract-utilities" Dec 03 00:32:26 crc kubenswrapper[4953]: I1203 00:32:26.464499 4953 state_mem.go:107] "Deleted CPUSet assignment" podUID="a398a67b-f052-465e-8b7f-3913a21f1adf" containerName="extract-utilities" Dec 03 00:32:26 crc kubenswrapper[4953]: I1203 00:32:26.464627 4953 memory_manager.go:354] "RemoveStaleState removing state" podUID="c9f7486f-c563-434d-b202-9211da3efe5c" containerName="registry-server" Dec 03 00:32:26 crc kubenswrapper[4953]: I1203 00:32:26.464641 4953 memory_manager.go:354] "RemoveStaleState removing state" podUID="a398a67b-f052-465e-8b7f-3913a21f1adf" containerName="registry-server" Dec 03 00:32:26 crc kubenswrapper[4953]: I1203 00:32:26.464662 4953 memory_manager.go:354] "RemoveStaleState removing state" podUID="2b0995a0-793c-4aac-9d8d-bc0dd4604b29" containerName="docker-build" Dec 03 00:32:26 crc kubenswrapper[4953]: I1203 00:32:26.465430 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/prometheus-webhook-snmp-1-build" Dec 03 00:32:26 crc kubenswrapper[4953]: I1203 00:32:26.469366 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"prometheus-webhook-snmp-1-sys-config" Dec 03 00:32:26 crc kubenswrapper[4953]: I1203 00:32:26.469413 4953 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"builder-dockercfg-9prwd" Dec 03 00:32:26 crc kubenswrapper[4953]: I1203 00:32:26.469365 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"prometheus-webhook-snmp-1-ca" Dec 03 00:32:26 crc kubenswrapper[4953]: I1203 00:32:26.471052 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"prometheus-webhook-snmp-1-global-ca" Dec 03 00:32:26 crc kubenswrapper[4953]: I1203 00:32:26.489759 4953 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/prometheus-webhook-snmp-1-build"] Dec 03 00:32:26 crc kubenswrapper[4953]: I1203 00:32:26.597419 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/5dbca6cf-05f2-487a-9d13-c5b9cc9c8740-container-storage-run\") pod \"prometheus-webhook-snmp-1-build\" (UID: \"5dbca6cf-05f2-487a-9d13-c5b9cc9c8740\") " pod="service-telemetry/prometheus-webhook-snmp-1-build" Dec 03 00:32:26 crc kubenswrapper[4953]: I1203 00:32:26.597520 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sfv75\" (UniqueName: \"kubernetes.io/projected/5dbca6cf-05f2-487a-9d13-c5b9cc9c8740-kube-api-access-sfv75\") pod \"prometheus-webhook-snmp-1-build\" (UID: \"5dbca6cf-05f2-487a-9d13-c5b9cc9c8740\") " pod="service-telemetry/prometheus-webhook-snmp-1-build" Dec 03 00:32:26 crc kubenswrapper[4953]: I1203 00:32:26.597551 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/5dbca6cf-05f2-487a-9d13-c5b9cc9c8740-build-proxy-ca-bundles\") pod \"prometheus-webhook-snmp-1-build\" (UID: \"5dbca6cf-05f2-487a-9d13-c5b9cc9c8740\") " pod="service-telemetry/prometheus-webhook-snmp-1-build" Dec 03 00:32:26 crc kubenswrapper[4953]: I1203 00:32:26.597582 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/5dbca6cf-05f2-487a-9d13-c5b9cc9c8740-build-system-configs\") pod \"prometheus-webhook-snmp-1-build\" (UID: \"5dbca6cf-05f2-487a-9d13-c5b9cc9c8740\") " pod="service-telemetry/prometheus-webhook-snmp-1-build" Dec 03 00:32:26 crc kubenswrapper[4953]: I1203 00:32:26.597618 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/5dbca6cf-05f2-487a-9d13-c5b9cc9c8740-container-storage-root\") pod \"prometheus-webhook-snmp-1-build\" (UID: \"5dbca6cf-05f2-487a-9d13-c5b9cc9c8740\") " pod="service-telemetry/prometheus-webhook-snmp-1-build" Dec 03 00:32:26 crc kubenswrapper[4953]: I1203 00:32:26.597660 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/5dbca6cf-05f2-487a-9d13-c5b9cc9c8740-build-ca-bundles\") pod \"prometheus-webhook-snmp-1-build\" (UID: \"5dbca6cf-05f2-487a-9d13-c5b9cc9c8740\") " pod="service-telemetry/prometheus-webhook-snmp-1-build" Dec 03 00:32:26 crc kubenswrapper[4953]: I1203 00:32:26.597695 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/5dbca6cf-05f2-487a-9d13-c5b9cc9c8740-build-blob-cache\") pod \"prometheus-webhook-snmp-1-build\" (UID: \"5dbca6cf-05f2-487a-9d13-c5b9cc9c8740\") " pod="service-telemetry/prometheus-webhook-snmp-1-build" Dec 03 00:32:26 crc kubenswrapper[4953]: I1203 00:32:26.597729 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/5dbca6cf-05f2-487a-9d13-c5b9cc9c8740-node-pullsecrets\") pod \"prometheus-webhook-snmp-1-build\" (UID: \"5dbca6cf-05f2-487a-9d13-c5b9cc9c8740\") " pod="service-telemetry/prometheus-webhook-snmp-1-build" Dec 03 00:32:26 crc kubenswrapper[4953]: I1203 00:32:26.597766 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/5dbca6cf-05f2-487a-9d13-c5b9cc9c8740-buildworkdir\") pod \"prometheus-webhook-snmp-1-build\" (UID: \"5dbca6cf-05f2-487a-9d13-c5b9cc9c8740\") " pod="service-telemetry/prometheus-webhook-snmp-1-build" Dec 03 00:32:26 crc kubenswrapper[4953]: I1203 00:32:26.597806 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/5dbca6cf-05f2-487a-9d13-c5b9cc9c8740-buildcachedir\") pod \"prometheus-webhook-snmp-1-build\" (UID: \"5dbca6cf-05f2-487a-9d13-c5b9cc9c8740\") " pod="service-telemetry/prometheus-webhook-snmp-1-build" Dec 03 00:32:26 crc kubenswrapper[4953]: I1203 00:32:26.597835 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"builder-dockercfg-9prwd-pull\" (UniqueName: \"kubernetes.io/secret/5dbca6cf-05f2-487a-9d13-c5b9cc9c8740-builder-dockercfg-9prwd-pull\") pod \"prometheus-webhook-snmp-1-build\" (UID: \"5dbca6cf-05f2-487a-9d13-c5b9cc9c8740\") " pod="service-telemetry/prometheus-webhook-snmp-1-build" Dec 03 00:32:26 crc kubenswrapper[4953]: I1203 00:32:26.597873 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"builder-dockercfg-9prwd-push\" (UniqueName: \"kubernetes.io/secret/5dbca6cf-05f2-487a-9d13-c5b9cc9c8740-builder-dockercfg-9prwd-push\") pod \"prometheus-webhook-snmp-1-build\" (UID: \"5dbca6cf-05f2-487a-9d13-c5b9cc9c8740\") " pod="service-telemetry/prometheus-webhook-snmp-1-build" Dec 03 00:32:26 crc kubenswrapper[4953]: I1203 00:32:26.703068 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/5dbca6cf-05f2-487a-9d13-c5b9cc9c8740-buildcachedir\") pod \"prometheus-webhook-snmp-1-build\" (UID: \"5dbca6cf-05f2-487a-9d13-c5b9cc9c8740\") " pod="service-telemetry/prometheus-webhook-snmp-1-build" Dec 03 00:32:26 crc kubenswrapper[4953]: I1203 00:32:26.703156 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"builder-dockercfg-9prwd-pull\" (UniqueName: \"kubernetes.io/secret/5dbca6cf-05f2-487a-9d13-c5b9cc9c8740-builder-dockercfg-9prwd-pull\") pod \"prometheus-webhook-snmp-1-build\" (UID: \"5dbca6cf-05f2-487a-9d13-c5b9cc9c8740\") " pod="service-telemetry/prometheus-webhook-snmp-1-build" Dec 03 00:32:26 crc kubenswrapper[4953]: I1203 00:32:26.703202 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"builder-dockercfg-9prwd-push\" (UniqueName: \"kubernetes.io/secret/5dbca6cf-05f2-487a-9d13-c5b9cc9c8740-builder-dockercfg-9prwd-push\") pod \"prometheus-webhook-snmp-1-build\" (UID: \"5dbca6cf-05f2-487a-9d13-c5b9cc9c8740\") " pod="service-telemetry/prometheus-webhook-snmp-1-build" Dec 03 00:32:26 crc kubenswrapper[4953]: I1203 00:32:26.703252 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/5dbca6cf-05f2-487a-9d13-c5b9cc9c8740-container-storage-run\") pod \"prometheus-webhook-snmp-1-build\" (UID: \"5dbca6cf-05f2-487a-9d13-c5b9cc9c8740\") " pod="service-telemetry/prometheus-webhook-snmp-1-build" Dec 03 00:32:26 crc kubenswrapper[4953]: I1203 00:32:26.703285 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/5dbca6cf-05f2-487a-9d13-c5b9cc9c8740-build-proxy-ca-bundles\") pod \"prometheus-webhook-snmp-1-build\" (UID: \"5dbca6cf-05f2-487a-9d13-c5b9cc9c8740\") " pod="service-telemetry/prometheus-webhook-snmp-1-build" Dec 03 00:32:26 crc kubenswrapper[4953]: I1203 00:32:26.703312 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sfv75\" (UniqueName: \"kubernetes.io/projected/5dbca6cf-05f2-487a-9d13-c5b9cc9c8740-kube-api-access-sfv75\") pod \"prometheus-webhook-snmp-1-build\" (UID: \"5dbca6cf-05f2-487a-9d13-c5b9cc9c8740\") " pod="service-telemetry/prometheus-webhook-snmp-1-build" Dec 03 00:32:26 crc kubenswrapper[4953]: I1203 00:32:26.703333 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/5dbca6cf-05f2-487a-9d13-c5b9cc9c8740-build-system-configs\") pod \"prometheus-webhook-snmp-1-build\" (UID: \"5dbca6cf-05f2-487a-9d13-c5b9cc9c8740\") " pod="service-telemetry/prometheus-webhook-snmp-1-build" Dec 03 00:32:26 crc kubenswrapper[4953]: I1203 00:32:26.703389 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/5dbca6cf-05f2-487a-9d13-c5b9cc9c8740-container-storage-root\") pod \"prometheus-webhook-snmp-1-build\" (UID: \"5dbca6cf-05f2-487a-9d13-c5b9cc9c8740\") " pod="service-telemetry/prometheus-webhook-snmp-1-build" Dec 03 00:32:26 crc kubenswrapper[4953]: I1203 00:32:26.703427 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/5dbca6cf-05f2-487a-9d13-c5b9cc9c8740-build-ca-bundles\") pod \"prometheus-webhook-snmp-1-build\" (UID: \"5dbca6cf-05f2-487a-9d13-c5b9cc9c8740\") " pod="service-telemetry/prometheus-webhook-snmp-1-build" Dec 03 00:32:26 crc kubenswrapper[4953]: I1203 00:32:26.703465 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/5dbca6cf-05f2-487a-9d13-c5b9cc9c8740-build-blob-cache\") pod \"prometheus-webhook-snmp-1-build\" (UID: \"5dbca6cf-05f2-487a-9d13-c5b9cc9c8740\") " pod="service-telemetry/prometheus-webhook-snmp-1-build" Dec 03 00:32:26 crc kubenswrapper[4953]: I1203 00:32:26.703496 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/5dbca6cf-05f2-487a-9d13-c5b9cc9c8740-node-pullsecrets\") pod \"prometheus-webhook-snmp-1-build\" (UID: \"5dbca6cf-05f2-487a-9d13-c5b9cc9c8740\") " pod="service-telemetry/prometheus-webhook-snmp-1-build" Dec 03 00:32:26 crc kubenswrapper[4953]: I1203 00:32:26.703526 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/5dbca6cf-05f2-487a-9d13-c5b9cc9c8740-buildworkdir\") pod \"prometheus-webhook-snmp-1-build\" (UID: \"5dbca6cf-05f2-487a-9d13-c5b9cc9c8740\") " pod="service-telemetry/prometheus-webhook-snmp-1-build" Dec 03 00:32:26 crc kubenswrapper[4953]: I1203 00:32:26.704389 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/5dbca6cf-05f2-487a-9d13-c5b9cc9c8740-buildworkdir\") pod \"prometheus-webhook-snmp-1-build\" (UID: \"5dbca6cf-05f2-487a-9d13-c5b9cc9c8740\") " pod="service-telemetry/prometheus-webhook-snmp-1-build" Dec 03 00:32:26 crc kubenswrapper[4953]: I1203 00:32:26.704946 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/5dbca6cf-05f2-487a-9d13-c5b9cc9c8740-build-proxy-ca-bundles\") pod \"prometheus-webhook-snmp-1-build\" (UID: \"5dbca6cf-05f2-487a-9d13-c5b9cc9c8740\") " pod="service-telemetry/prometheus-webhook-snmp-1-build" Dec 03 00:32:26 crc kubenswrapper[4953]: I1203 00:32:26.705060 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/5dbca6cf-05f2-487a-9d13-c5b9cc9c8740-buildcachedir\") pod \"prometheus-webhook-snmp-1-build\" (UID: \"5dbca6cf-05f2-487a-9d13-c5b9cc9c8740\") " pod="service-telemetry/prometheus-webhook-snmp-1-build" Dec 03 00:32:26 crc kubenswrapper[4953]: I1203 00:32:26.705306 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/5dbca6cf-05f2-487a-9d13-c5b9cc9c8740-build-system-configs\") pod \"prometheus-webhook-snmp-1-build\" (UID: \"5dbca6cf-05f2-487a-9d13-c5b9cc9c8740\") " pod="service-telemetry/prometheus-webhook-snmp-1-build" Dec 03 00:32:26 crc kubenswrapper[4953]: I1203 00:32:26.705457 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/5dbca6cf-05f2-487a-9d13-c5b9cc9c8740-container-storage-run\") pod \"prometheus-webhook-snmp-1-build\" (UID: \"5dbca6cf-05f2-487a-9d13-c5b9cc9c8740\") " pod="service-telemetry/prometheus-webhook-snmp-1-build" Dec 03 00:32:26 crc kubenswrapper[4953]: I1203 00:32:26.705572 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/5dbca6cf-05f2-487a-9d13-c5b9cc9c8740-node-pullsecrets\") pod \"prometheus-webhook-snmp-1-build\" (UID: \"5dbca6cf-05f2-487a-9d13-c5b9cc9c8740\") " pod="service-telemetry/prometheus-webhook-snmp-1-build" Dec 03 00:32:26 crc kubenswrapper[4953]: I1203 00:32:26.705593 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/5dbca6cf-05f2-487a-9d13-c5b9cc9c8740-build-blob-cache\") pod \"prometheus-webhook-snmp-1-build\" (UID: \"5dbca6cf-05f2-487a-9d13-c5b9cc9c8740\") " pod="service-telemetry/prometheus-webhook-snmp-1-build" Dec 03 00:32:26 crc kubenswrapper[4953]: I1203 00:32:26.706236 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/5dbca6cf-05f2-487a-9d13-c5b9cc9c8740-container-storage-root\") pod \"prometheus-webhook-snmp-1-build\" (UID: \"5dbca6cf-05f2-487a-9d13-c5b9cc9c8740\") " pod="service-telemetry/prometheus-webhook-snmp-1-build" Dec 03 00:32:26 crc kubenswrapper[4953]: I1203 00:32:26.706377 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/5dbca6cf-05f2-487a-9d13-c5b9cc9c8740-build-ca-bundles\") pod \"prometheus-webhook-snmp-1-build\" (UID: \"5dbca6cf-05f2-487a-9d13-c5b9cc9c8740\") " pod="service-telemetry/prometheus-webhook-snmp-1-build" Dec 03 00:32:26 crc kubenswrapper[4953]: I1203 00:32:26.712352 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"builder-dockercfg-9prwd-push\" (UniqueName: \"kubernetes.io/secret/5dbca6cf-05f2-487a-9d13-c5b9cc9c8740-builder-dockercfg-9prwd-push\") pod \"prometheus-webhook-snmp-1-build\" (UID: \"5dbca6cf-05f2-487a-9d13-c5b9cc9c8740\") " pod="service-telemetry/prometheus-webhook-snmp-1-build" Dec 03 00:32:26 crc kubenswrapper[4953]: I1203 00:32:26.715389 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"builder-dockercfg-9prwd-pull\" (UniqueName: \"kubernetes.io/secret/5dbca6cf-05f2-487a-9d13-c5b9cc9c8740-builder-dockercfg-9prwd-pull\") pod \"prometheus-webhook-snmp-1-build\" (UID: \"5dbca6cf-05f2-487a-9d13-c5b9cc9c8740\") " pod="service-telemetry/prometheus-webhook-snmp-1-build" Dec 03 00:32:26 crc kubenswrapper[4953]: I1203 00:32:26.724926 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sfv75\" (UniqueName: \"kubernetes.io/projected/5dbca6cf-05f2-487a-9d13-c5b9cc9c8740-kube-api-access-sfv75\") pod \"prometheus-webhook-snmp-1-build\" (UID: \"5dbca6cf-05f2-487a-9d13-c5b9cc9c8740\") " pod="service-telemetry/prometheus-webhook-snmp-1-build" Dec 03 00:32:26 crc kubenswrapper[4953]: I1203 00:32:26.788873 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/prometheus-webhook-snmp-1-build" Dec 03 00:32:27 crc kubenswrapper[4953]: I1203 00:32:27.056316 4953 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/prometheus-webhook-snmp-1-build"] Dec 03 00:32:27 crc kubenswrapper[4953]: I1203 00:32:27.625002 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/prometheus-webhook-snmp-1-build" event={"ID":"5dbca6cf-05f2-487a-9d13-c5b9cc9c8740","Type":"ContainerStarted","Data":"c9de930864a423f4542d06ed53ba3dbbccc4172c79ae8ebc21e7c73bc459e555"} Dec 03 00:32:28 crc kubenswrapper[4953]: I1203 00:32:28.637748 4953 generic.go:334] "Generic (PLEG): container finished" podID="5dbca6cf-05f2-487a-9d13-c5b9cc9c8740" containerID="3b09a77d64394775167f15c217c17552cc7a359d1353dc8e03bfd6508d53fcc2" exitCode=0 Dec 03 00:32:28 crc kubenswrapper[4953]: I1203 00:32:28.637929 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/prometheus-webhook-snmp-1-build" event={"ID":"5dbca6cf-05f2-487a-9d13-c5b9cc9c8740","Type":"ContainerDied","Data":"3b09a77d64394775167f15c217c17552cc7a359d1353dc8e03bfd6508d53fcc2"} Dec 03 00:32:29 crc kubenswrapper[4953]: I1203 00:32:29.645907 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/prometheus-webhook-snmp-1-build" event={"ID":"5dbca6cf-05f2-487a-9d13-c5b9cc9c8740","Type":"ContainerStarted","Data":"c906e26ca6cb9d991a923cfa26b3f1dd62b257abb080b19531ba08e1bc1d0dd4"} Dec 03 00:32:29 crc kubenswrapper[4953]: I1203 00:32:29.691693 4953 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="service-telemetry/prometheus-webhook-snmp-1-build" podStartSLOduration=3.691667824 podStartE2EDuration="3.691667824s" podCreationTimestamp="2025-12-03 00:32:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 00:32:29.677452955 +0000 UTC m=+1538.661733778" watchObservedRunningTime="2025-12-03 00:32:29.691667824 +0000 UTC m=+1538.675948647" Dec 03 00:32:36 crc kubenswrapper[4953]: I1203 00:32:36.517905 4953 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["service-telemetry/prometheus-webhook-snmp-1-build"] Dec 03 00:32:36 crc kubenswrapper[4953]: I1203 00:32:36.519351 4953 kuberuntime_container.go:808] "Killing container with a grace period" pod="service-telemetry/prometheus-webhook-snmp-1-build" podUID="5dbca6cf-05f2-487a-9d13-c5b9cc9c8740" containerName="docker-build" containerID="cri-o://c906e26ca6cb9d991a923cfa26b3f1dd62b257abb080b19531ba08e1bc1d0dd4" gracePeriod=30 Dec 03 00:32:36 crc kubenswrapper[4953]: I1203 00:32:36.700120 4953 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_prometheus-webhook-snmp-1-build_5dbca6cf-05f2-487a-9d13-c5b9cc9c8740/docker-build/0.log" Dec 03 00:32:36 crc kubenswrapper[4953]: I1203 00:32:36.701754 4953 generic.go:334] "Generic (PLEG): container finished" podID="5dbca6cf-05f2-487a-9d13-c5b9cc9c8740" containerID="c906e26ca6cb9d991a923cfa26b3f1dd62b257abb080b19531ba08e1bc1d0dd4" exitCode=1 Dec 03 00:32:36 crc kubenswrapper[4953]: I1203 00:32:36.701813 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/prometheus-webhook-snmp-1-build" event={"ID":"5dbca6cf-05f2-487a-9d13-c5b9cc9c8740","Type":"ContainerDied","Data":"c906e26ca6cb9d991a923cfa26b3f1dd62b257abb080b19531ba08e1bc1d0dd4"} Dec 03 00:32:36 crc kubenswrapper[4953]: I1203 00:32:36.919840 4953 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_prometheus-webhook-snmp-1-build_5dbca6cf-05f2-487a-9d13-c5b9cc9c8740/docker-build/0.log" Dec 03 00:32:36 crc kubenswrapper[4953]: I1203 00:32:36.922540 4953 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/prometheus-webhook-snmp-1-build" Dec 03 00:32:37 crc kubenswrapper[4953]: I1203 00:32:37.063786 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/5dbca6cf-05f2-487a-9d13-c5b9cc9c8740-container-storage-run\") pod \"5dbca6cf-05f2-487a-9d13-c5b9cc9c8740\" (UID: \"5dbca6cf-05f2-487a-9d13-c5b9cc9c8740\") " Dec 03 00:32:37 crc kubenswrapper[4953]: I1203 00:32:37.063905 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/5dbca6cf-05f2-487a-9d13-c5b9cc9c8740-node-pullsecrets\") pod \"5dbca6cf-05f2-487a-9d13-c5b9cc9c8740\" (UID: \"5dbca6cf-05f2-487a-9d13-c5b9cc9c8740\") " Dec 03 00:32:37 crc kubenswrapper[4953]: I1203 00:32:37.063939 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/5dbca6cf-05f2-487a-9d13-c5b9cc9c8740-build-proxy-ca-bundles\") pod \"5dbca6cf-05f2-487a-9d13-c5b9cc9c8740\" (UID: \"5dbca6cf-05f2-487a-9d13-c5b9cc9c8740\") " Dec 03 00:32:37 crc kubenswrapper[4953]: I1203 00:32:37.063964 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sfv75\" (UniqueName: \"kubernetes.io/projected/5dbca6cf-05f2-487a-9d13-c5b9cc9c8740-kube-api-access-sfv75\") pod \"5dbca6cf-05f2-487a-9d13-c5b9cc9c8740\" (UID: \"5dbca6cf-05f2-487a-9d13-c5b9cc9c8740\") " Dec 03 00:32:37 crc kubenswrapper[4953]: I1203 00:32:37.064004 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/5dbca6cf-05f2-487a-9d13-c5b9cc9c8740-container-storage-root\") pod \"5dbca6cf-05f2-487a-9d13-c5b9cc9c8740\" (UID: \"5dbca6cf-05f2-487a-9d13-c5b9cc9c8740\") " Dec 03 00:32:37 crc kubenswrapper[4953]: I1203 00:32:37.064028 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/5dbca6cf-05f2-487a-9d13-c5b9cc9c8740-buildworkdir\") pod \"5dbca6cf-05f2-487a-9d13-c5b9cc9c8740\" (UID: \"5dbca6cf-05f2-487a-9d13-c5b9cc9c8740\") " Dec 03 00:32:37 crc kubenswrapper[4953]: I1203 00:32:37.064013 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5dbca6cf-05f2-487a-9d13-c5b9cc9c8740-node-pullsecrets" (OuterVolumeSpecName: "node-pullsecrets") pod "5dbca6cf-05f2-487a-9d13-c5b9cc9c8740" (UID: "5dbca6cf-05f2-487a-9d13-c5b9cc9c8740"). InnerVolumeSpecName "node-pullsecrets". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 00:32:37 crc kubenswrapper[4953]: I1203 00:32:37.064228 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/5dbca6cf-05f2-487a-9d13-c5b9cc9c8740-build-blob-cache\") pod \"5dbca6cf-05f2-487a-9d13-c5b9cc9c8740\" (UID: \"5dbca6cf-05f2-487a-9d13-c5b9cc9c8740\") " Dec 03 00:32:37 crc kubenswrapper[4953]: I1203 00:32:37.064305 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/5dbca6cf-05f2-487a-9d13-c5b9cc9c8740-buildcachedir\") pod \"5dbca6cf-05f2-487a-9d13-c5b9cc9c8740\" (UID: \"5dbca6cf-05f2-487a-9d13-c5b9cc9c8740\") " Dec 03 00:32:37 crc kubenswrapper[4953]: I1203 00:32:37.064375 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/5dbca6cf-05f2-487a-9d13-c5b9cc9c8740-build-ca-bundles\") pod \"5dbca6cf-05f2-487a-9d13-c5b9cc9c8740\" (UID: \"5dbca6cf-05f2-487a-9d13-c5b9cc9c8740\") " Dec 03 00:32:37 crc kubenswrapper[4953]: I1203 00:32:37.064442 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"builder-dockercfg-9prwd-push\" (UniqueName: \"kubernetes.io/secret/5dbca6cf-05f2-487a-9d13-c5b9cc9c8740-builder-dockercfg-9prwd-push\") pod \"5dbca6cf-05f2-487a-9d13-c5b9cc9c8740\" (UID: \"5dbca6cf-05f2-487a-9d13-c5b9cc9c8740\") " Dec 03 00:32:37 crc kubenswrapper[4953]: I1203 00:32:37.064466 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"builder-dockercfg-9prwd-pull\" (UniqueName: \"kubernetes.io/secret/5dbca6cf-05f2-487a-9d13-c5b9cc9c8740-builder-dockercfg-9prwd-pull\") pod \"5dbca6cf-05f2-487a-9d13-c5b9cc9c8740\" (UID: \"5dbca6cf-05f2-487a-9d13-c5b9cc9c8740\") " Dec 03 00:32:37 crc kubenswrapper[4953]: I1203 00:32:37.064464 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5dbca6cf-05f2-487a-9d13-c5b9cc9c8740-buildcachedir" (OuterVolumeSpecName: "buildcachedir") pod "5dbca6cf-05f2-487a-9d13-c5b9cc9c8740" (UID: "5dbca6cf-05f2-487a-9d13-c5b9cc9c8740"). InnerVolumeSpecName "buildcachedir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 00:32:37 crc kubenswrapper[4953]: I1203 00:32:37.064530 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/5dbca6cf-05f2-487a-9d13-c5b9cc9c8740-build-system-configs\") pod \"5dbca6cf-05f2-487a-9d13-c5b9cc9c8740\" (UID: \"5dbca6cf-05f2-487a-9d13-c5b9cc9c8740\") " Dec 03 00:32:37 crc kubenswrapper[4953]: I1203 00:32:37.065204 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5dbca6cf-05f2-487a-9d13-c5b9cc9c8740-buildworkdir" (OuterVolumeSpecName: "buildworkdir") pod "5dbca6cf-05f2-487a-9d13-c5b9cc9c8740" (UID: "5dbca6cf-05f2-487a-9d13-c5b9cc9c8740"). InnerVolumeSpecName "buildworkdir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 00:32:37 crc kubenswrapper[4953]: I1203 00:32:37.065244 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5dbca6cf-05f2-487a-9d13-c5b9cc9c8740-build-proxy-ca-bundles" (OuterVolumeSpecName: "build-proxy-ca-bundles") pod "5dbca6cf-05f2-487a-9d13-c5b9cc9c8740" (UID: "5dbca6cf-05f2-487a-9d13-c5b9cc9c8740"). InnerVolumeSpecName "build-proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:32:37 crc kubenswrapper[4953]: I1203 00:32:37.065492 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5dbca6cf-05f2-487a-9d13-c5b9cc9c8740-container-storage-run" (OuterVolumeSpecName: "container-storage-run") pod "5dbca6cf-05f2-487a-9d13-c5b9cc9c8740" (UID: "5dbca6cf-05f2-487a-9d13-c5b9cc9c8740"). InnerVolumeSpecName "container-storage-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 00:32:37 crc kubenswrapper[4953]: I1203 00:32:37.065682 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5dbca6cf-05f2-487a-9d13-c5b9cc9c8740-build-ca-bundles" (OuterVolumeSpecName: "build-ca-bundles") pod "5dbca6cf-05f2-487a-9d13-c5b9cc9c8740" (UID: "5dbca6cf-05f2-487a-9d13-c5b9cc9c8740"). InnerVolumeSpecName "build-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:32:37 crc kubenswrapper[4953]: I1203 00:32:37.065796 4953 reconciler_common.go:293] "Volume detached for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/5dbca6cf-05f2-487a-9d13-c5b9cc9c8740-node-pullsecrets\") on node \"crc\" DevicePath \"\"" Dec 03 00:32:37 crc kubenswrapper[4953]: I1203 00:32:37.065820 4953 reconciler_common.go:293] "Volume detached for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/5dbca6cf-05f2-487a-9d13-c5b9cc9c8740-build-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Dec 03 00:32:37 crc kubenswrapper[4953]: I1203 00:32:37.065837 4953 reconciler_common.go:293] "Volume detached for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/5dbca6cf-05f2-487a-9d13-c5b9cc9c8740-buildworkdir\") on node \"crc\" DevicePath \"\"" Dec 03 00:32:37 crc kubenswrapper[4953]: I1203 00:32:37.065849 4953 reconciler_common.go:293] "Volume detached for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/5dbca6cf-05f2-487a-9d13-c5b9cc9c8740-buildcachedir\") on node \"crc\" DevicePath \"\"" Dec 03 00:32:37 crc kubenswrapper[4953]: I1203 00:32:37.066063 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5dbca6cf-05f2-487a-9d13-c5b9cc9c8740-build-system-configs" (OuterVolumeSpecName: "build-system-configs") pod "5dbca6cf-05f2-487a-9d13-c5b9cc9c8740" (UID: "5dbca6cf-05f2-487a-9d13-c5b9cc9c8740"). InnerVolumeSpecName "build-system-configs". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:32:37 crc kubenswrapper[4953]: I1203 00:32:37.072664 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5dbca6cf-05f2-487a-9d13-c5b9cc9c8740-kube-api-access-sfv75" (OuterVolumeSpecName: "kube-api-access-sfv75") pod "5dbca6cf-05f2-487a-9d13-c5b9cc9c8740" (UID: "5dbca6cf-05f2-487a-9d13-c5b9cc9c8740"). InnerVolumeSpecName "kube-api-access-sfv75". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:32:37 crc kubenswrapper[4953]: I1203 00:32:37.086328 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5dbca6cf-05f2-487a-9d13-c5b9cc9c8740-builder-dockercfg-9prwd-pull" (OuterVolumeSpecName: "builder-dockercfg-9prwd-pull") pod "5dbca6cf-05f2-487a-9d13-c5b9cc9c8740" (UID: "5dbca6cf-05f2-487a-9d13-c5b9cc9c8740"). InnerVolumeSpecName "builder-dockercfg-9prwd-pull". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:32:37 crc kubenswrapper[4953]: I1203 00:32:37.086409 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5dbca6cf-05f2-487a-9d13-c5b9cc9c8740-builder-dockercfg-9prwd-push" (OuterVolumeSpecName: "builder-dockercfg-9prwd-push") pod "5dbca6cf-05f2-487a-9d13-c5b9cc9c8740" (UID: "5dbca6cf-05f2-487a-9d13-c5b9cc9c8740"). InnerVolumeSpecName "builder-dockercfg-9prwd-push". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:32:37 crc kubenswrapper[4953]: I1203 00:32:37.132121 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5dbca6cf-05f2-487a-9d13-c5b9cc9c8740-build-blob-cache" (OuterVolumeSpecName: "build-blob-cache") pod "5dbca6cf-05f2-487a-9d13-c5b9cc9c8740" (UID: "5dbca6cf-05f2-487a-9d13-c5b9cc9c8740"). InnerVolumeSpecName "build-blob-cache". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 00:32:37 crc kubenswrapper[4953]: I1203 00:32:37.167442 4953 reconciler_common.go:293] "Volume detached for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/5dbca6cf-05f2-487a-9d13-c5b9cc9c8740-build-blob-cache\") on node \"crc\" DevicePath \"\"" Dec 03 00:32:37 crc kubenswrapper[4953]: I1203 00:32:37.167495 4953 reconciler_common.go:293] "Volume detached for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/5dbca6cf-05f2-487a-9d13-c5b9cc9c8740-build-ca-bundles\") on node \"crc\" DevicePath \"\"" Dec 03 00:32:37 crc kubenswrapper[4953]: I1203 00:32:37.167506 4953 reconciler_common.go:293] "Volume detached for volume \"builder-dockercfg-9prwd-push\" (UniqueName: \"kubernetes.io/secret/5dbca6cf-05f2-487a-9d13-c5b9cc9c8740-builder-dockercfg-9prwd-push\") on node \"crc\" DevicePath \"\"" Dec 03 00:32:37 crc kubenswrapper[4953]: I1203 00:32:37.167521 4953 reconciler_common.go:293] "Volume detached for volume \"builder-dockercfg-9prwd-pull\" (UniqueName: \"kubernetes.io/secret/5dbca6cf-05f2-487a-9d13-c5b9cc9c8740-builder-dockercfg-9prwd-pull\") on node \"crc\" DevicePath \"\"" Dec 03 00:32:37 crc kubenswrapper[4953]: I1203 00:32:37.167532 4953 reconciler_common.go:293] "Volume detached for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/5dbca6cf-05f2-487a-9d13-c5b9cc9c8740-build-system-configs\") on node \"crc\" DevicePath \"\"" Dec 03 00:32:37 crc kubenswrapper[4953]: I1203 00:32:37.167541 4953 reconciler_common.go:293] "Volume detached for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/5dbca6cf-05f2-487a-9d13-c5b9cc9c8740-container-storage-run\") on node \"crc\" DevicePath \"\"" Dec 03 00:32:37 crc kubenswrapper[4953]: I1203 00:32:37.167549 4953 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sfv75\" (UniqueName: \"kubernetes.io/projected/5dbca6cf-05f2-487a-9d13-c5b9cc9c8740-kube-api-access-sfv75\") on node \"crc\" DevicePath \"\"" Dec 03 00:32:37 crc kubenswrapper[4953]: I1203 00:32:37.408857 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5dbca6cf-05f2-487a-9d13-c5b9cc9c8740-container-storage-root" (OuterVolumeSpecName: "container-storage-root") pod "5dbca6cf-05f2-487a-9d13-c5b9cc9c8740" (UID: "5dbca6cf-05f2-487a-9d13-c5b9cc9c8740"). InnerVolumeSpecName "container-storage-root". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 00:32:37 crc kubenswrapper[4953]: I1203 00:32:37.472248 4953 reconciler_common.go:293] "Volume detached for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/5dbca6cf-05f2-487a-9d13-c5b9cc9c8740-container-storage-root\") on node \"crc\" DevicePath \"\"" Dec 03 00:32:37 crc kubenswrapper[4953]: I1203 00:32:37.727939 4953 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_prometheus-webhook-snmp-1-build_5dbca6cf-05f2-487a-9d13-c5b9cc9c8740/docker-build/0.log" Dec 03 00:32:37 crc kubenswrapper[4953]: I1203 00:32:37.728767 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/prometheus-webhook-snmp-1-build" event={"ID":"5dbca6cf-05f2-487a-9d13-c5b9cc9c8740","Type":"ContainerDied","Data":"c9de930864a423f4542d06ed53ba3dbbccc4172c79ae8ebc21e7c73bc459e555"} Dec 03 00:32:37 crc kubenswrapper[4953]: I1203 00:32:37.728868 4953 scope.go:117] "RemoveContainer" containerID="c906e26ca6cb9d991a923cfa26b3f1dd62b257abb080b19531ba08e1bc1d0dd4" Dec 03 00:32:37 crc kubenswrapper[4953]: I1203 00:32:37.728935 4953 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/prometheus-webhook-snmp-1-build" Dec 03 00:32:37 crc kubenswrapper[4953]: I1203 00:32:37.787957 4953 scope.go:117] "RemoveContainer" containerID="3b09a77d64394775167f15c217c17552cc7a359d1353dc8e03bfd6508d53fcc2" Dec 03 00:32:37 crc kubenswrapper[4953]: I1203 00:32:37.818128 4953 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["service-telemetry/prometheus-webhook-snmp-1-build"] Dec 03 00:32:37 crc kubenswrapper[4953]: I1203 00:32:37.823322 4953 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["service-telemetry/prometheus-webhook-snmp-1-build"] Dec 03 00:32:38 crc kubenswrapper[4953]: I1203 00:32:38.131040 4953 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["service-telemetry/prometheus-webhook-snmp-2-build"] Dec 03 00:32:38 crc kubenswrapper[4953]: E1203 00:32:38.131893 4953 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5dbca6cf-05f2-487a-9d13-c5b9cc9c8740" containerName="docker-build" Dec 03 00:32:38 crc kubenswrapper[4953]: I1203 00:32:38.131916 4953 state_mem.go:107] "Deleted CPUSet assignment" podUID="5dbca6cf-05f2-487a-9d13-c5b9cc9c8740" containerName="docker-build" Dec 03 00:32:38 crc kubenswrapper[4953]: E1203 00:32:38.131943 4953 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5dbca6cf-05f2-487a-9d13-c5b9cc9c8740" containerName="manage-dockerfile" Dec 03 00:32:38 crc kubenswrapper[4953]: I1203 00:32:38.131951 4953 state_mem.go:107] "Deleted CPUSet assignment" podUID="5dbca6cf-05f2-487a-9d13-c5b9cc9c8740" containerName="manage-dockerfile" Dec 03 00:32:38 crc kubenswrapper[4953]: I1203 00:32:38.132072 4953 memory_manager.go:354] "RemoveStaleState removing state" podUID="5dbca6cf-05f2-487a-9d13-c5b9cc9c8740" containerName="docker-build" Dec 03 00:32:38 crc kubenswrapper[4953]: I1203 00:32:38.133079 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/prometheus-webhook-snmp-2-build" Dec 03 00:32:38 crc kubenswrapper[4953]: I1203 00:32:38.135617 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"prometheus-webhook-snmp-2-ca" Dec 03 00:32:38 crc kubenswrapper[4953]: I1203 00:32:38.135641 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"prometheus-webhook-snmp-2-global-ca" Dec 03 00:32:38 crc kubenswrapper[4953]: I1203 00:32:38.135830 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"prometheus-webhook-snmp-2-sys-config" Dec 03 00:32:38 crc kubenswrapper[4953]: I1203 00:32:38.136269 4953 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"builder-dockercfg-9prwd" Dec 03 00:32:38 crc kubenswrapper[4953]: I1203 00:32:38.154260 4953 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/prometheus-webhook-snmp-2-build"] Dec 03 00:32:38 crc kubenswrapper[4953]: I1203 00:32:38.297535 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"builder-dockercfg-9prwd-push\" (UniqueName: \"kubernetes.io/secret/dfd0b89d-949b-459e-ad7c-0a2ef8aa3988-builder-dockercfg-9prwd-push\") pod \"prometheus-webhook-snmp-2-build\" (UID: \"dfd0b89d-949b-459e-ad7c-0a2ef8aa3988\") " pod="service-telemetry/prometheus-webhook-snmp-2-build" Dec 03 00:32:38 crc kubenswrapper[4953]: I1203 00:32:38.297625 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/dfd0b89d-949b-459e-ad7c-0a2ef8aa3988-buildcachedir\") pod \"prometheus-webhook-snmp-2-build\" (UID: \"dfd0b89d-949b-459e-ad7c-0a2ef8aa3988\") " pod="service-telemetry/prometheus-webhook-snmp-2-build" Dec 03 00:32:38 crc kubenswrapper[4953]: I1203 00:32:38.297667 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/dfd0b89d-949b-459e-ad7c-0a2ef8aa3988-build-system-configs\") pod \"prometheus-webhook-snmp-2-build\" (UID: \"dfd0b89d-949b-459e-ad7c-0a2ef8aa3988\") " pod="service-telemetry/prometheus-webhook-snmp-2-build" Dec 03 00:32:38 crc kubenswrapper[4953]: I1203 00:32:38.297865 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/dfd0b89d-949b-459e-ad7c-0a2ef8aa3988-container-storage-run\") pod \"prometheus-webhook-snmp-2-build\" (UID: \"dfd0b89d-949b-459e-ad7c-0a2ef8aa3988\") " pod="service-telemetry/prometheus-webhook-snmp-2-build" Dec 03 00:32:38 crc kubenswrapper[4953]: I1203 00:32:38.297926 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9rlpp\" (UniqueName: \"kubernetes.io/projected/dfd0b89d-949b-459e-ad7c-0a2ef8aa3988-kube-api-access-9rlpp\") pod \"prometheus-webhook-snmp-2-build\" (UID: \"dfd0b89d-949b-459e-ad7c-0a2ef8aa3988\") " pod="service-telemetry/prometheus-webhook-snmp-2-build" Dec 03 00:32:38 crc kubenswrapper[4953]: I1203 00:32:38.298134 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/dfd0b89d-949b-459e-ad7c-0a2ef8aa3988-build-blob-cache\") pod \"prometheus-webhook-snmp-2-build\" (UID: \"dfd0b89d-949b-459e-ad7c-0a2ef8aa3988\") " pod="service-telemetry/prometheus-webhook-snmp-2-build" Dec 03 00:32:38 crc kubenswrapper[4953]: I1203 00:32:38.298256 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/dfd0b89d-949b-459e-ad7c-0a2ef8aa3988-buildworkdir\") pod \"prometheus-webhook-snmp-2-build\" (UID: \"dfd0b89d-949b-459e-ad7c-0a2ef8aa3988\") " pod="service-telemetry/prometheus-webhook-snmp-2-build" Dec 03 00:32:38 crc kubenswrapper[4953]: I1203 00:32:38.298373 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/dfd0b89d-949b-459e-ad7c-0a2ef8aa3988-container-storage-root\") pod \"prometheus-webhook-snmp-2-build\" (UID: \"dfd0b89d-949b-459e-ad7c-0a2ef8aa3988\") " pod="service-telemetry/prometheus-webhook-snmp-2-build" Dec 03 00:32:38 crc kubenswrapper[4953]: I1203 00:32:38.298419 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/dfd0b89d-949b-459e-ad7c-0a2ef8aa3988-build-proxy-ca-bundles\") pod \"prometheus-webhook-snmp-2-build\" (UID: \"dfd0b89d-949b-459e-ad7c-0a2ef8aa3988\") " pod="service-telemetry/prometheus-webhook-snmp-2-build" Dec 03 00:32:38 crc kubenswrapper[4953]: I1203 00:32:38.298447 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/dfd0b89d-949b-459e-ad7c-0a2ef8aa3988-build-ca-bundles\") pod \"prometheus-webhook-snmp-2-build\" (UID: \"dfd0b89d-949b-459e-ad7c-0a2ef8aa3988\") " pod="service-telemetry/prometheus-webhook-snmp-2-build" Dec 03 00:32:38 crc kubenswrapper[4953]: I1203 00:32:38.298505 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"builder-dockercfg-9prwd-pull\" (UniqueName: \"kubernetes.io/secret/dfd0b89d-949b-459e-ad7c-0a2ef8aa3988-builder-dockercfg-9prwd-pull\") pod \"prometheus-webhook-snmp-2-build\" (UID: \"dfd0b89d-949b-459e-ad7c-0a2ef8aa3988\") " pod="service-telemetry/prometheus-webhook-snmp-2-build" Dec 03 00:32:38 crc kubenswrapper[4953]: I1203 00:32:38.298537 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/dfd0b89d-949b-459e-ad7c-0a2ef8aa3988-node-pullsecrets\") pod \"prometheus-webhook-snmp-2-build\" (UID: \"dfd0b89d-949b-459e-ad7c-0a2ef8aa3988\") " pod="service-telemetry/prometheus-webhook-snmp-2-build" Dec 03 00:32:38 crc kubenswrapper[4953]: I1203 00:32:38.400119 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/dfd0b89d-949b-459e-ad7c-0a2ef8aa3988-buildcachedir\") pod \"prometheus-webhook-snmp-2-build\" (UID: \"dfd0b89d-949b-459e-ad7c-0a2ef8aa3988\") " pod="service-telemetry/prometheus-webhook-snmp-2-build" Dec 03 00:32:38 crc kubenswrapper[4953]: I1203 00:32:38.400189 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/dfd0b89d-949b-459e-ad7c-0a2ef8aa3988-build-system-configs\") pod \"prometheus-webhook-snmp-2-build\" (UID: \"dfd0b89d-949b-459e-ad7c-0a2ef8aa3988\") " pod="service-telemetry/prometheus-webhook-snmp-2-build" Dec 03 00:32:38 crc kubenswrapper[4953]: I1203 00:32:38.400220 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/dfd0b89d-949b-459e-ad7c-0a2ef8aa3988-container-storage-run\") pod \"prometheus-webhook-snmp-2-build\" (UID: \"dfd0b89d-949b-459e-ad7c-0a2ef8aa3988\") " pod="service-telemetry/prometheus-webhook-snmp-2-build" Dec 03 00:32:38 crc kubenswrapper[4953]: I1203 00:32:38.400269 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9rlpp\" (UniqueName: \"kubernetes.io/projected/dfd0b89d-949b-459e-ad7c-0a2ef8aa3988-kube-api-access-9rlpp\") pod \"prometheus-webhook-snmp-2-build\" (UID: \"dfd0b89d-949b-459e-ad7c-0a2ef8aa3988\") " pod="service-telemetry/prometheus-webhook-snmp-2-build" Dec 03 00:32:38 crc kubenswrapper[4953]: I1203 00:32:38.400272 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/dfd0b89d-949b-459e-ad7c-0a2ef8aa3988-buildcachedir\") pod \"prometheus-webhook-snmp-2-build\" (UID: \"dfd0b89d-949b-459e-ad7c-0a2ef8aa3988\") " pod="service-telemetry/prometheus-webhook-snmp-2-build" Dec 03 00:32:38 crc kubenswrapper[4953]: I1203 00:32:38.400308 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/dfd0b89d-949b-459e-ad7c-0a2ef8aa3988-build-blob-cache\") pod \"prometheus-webhook-snmp-2-build\" (UID: \"dfd0b89d-949b-459e-ad7c-0a2ef8aa3988\") " pod="service-telemetry/prometheus-webhook-snmp-2-build" Dec 03 00:32:38 crc kubenswrapper[4953]: I1203 00:32:38.400341 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/dfd0b89d-949b-459e-ad7c-0a2ef8aa3988-buildworkdir\") pod \"prometheus-webhook-snmp-2-build\" (UID: \"dfd0b89d-949b-459e-ad7c-0a2ef8aa3988\") " pod="service-telemetry/prometheus-webhook-snmp-2-build" Dec 03 00:32:38 crc kubenswrapper[4953]: I1203 00:32:38.400397 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/dfd0b89d-949b-459e-ad7c-0a2ef8aa3988-container-storage-root\") pod \"prometheus-webhook-snmp-2-build\" (UID: \"dfd0b89d-949b-459e-ad7c-0a2ef8aa3988\") " pod="service-telemetry/prometheus-webhook-snmp-2-build" Dec 03 00:32:38 crc kubenswrapper[4953]: I1203 00:32:38.400434 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/dfd0b89d-949b-459e-ad7c-0a2ef8aa3988-build-proxy-ca-bundles\") pod \"prometheus-webhook-snmp-2-build\" (UID: \"dfd0b89d-949b-459e-ad7c-0a2ef8aa3988\") " pod="service-telemetry/prometheus-webhook-snmp-2-build" Dec 03 00:32:38 crc kubenswrapper[4953]: I1203 00:32:38.400466 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/dfd0b89d-949b-459e-ad7c-0a2ef8aa3988-build-ca-bundles\") pod \"prometheus-webhook-snmp-2-build\" (UID: \"dfd0b89d-949b-459e-ad7c-0a2ef8aa3988\") " pod="service-telemetry/prometheus-webhook-snmp-2-build" Dec 03 00:32:38 crc kubenswrapper[4953]: I1203 00:32:38.400501 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"builder-dockercfg-9prwd-pull\" (UniqueName: \"kubernetes.io/secret/dfd0b89d-949b-459e-ad7c-0a2ef8aa3988-builder-dockercfg-9prwd-pull\") pod \"prometheus-webhook-snmp-2-build\" (UID: \"dfd0b89d-949b-459e-ad7c-0a2ef8aa3988\") " pod="service-telemetry/prometheus-webhook-snmp-2-build" Dec 03 00:32:38 crc kubenswrapper[4953]: I1203 00:32:38.400524 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/dfd0b89d-949b-459e-ad7c-0a2ef8aa3988-node-pullsecrets\") pod \"prometheus-webhook-snmp-2-build\" (UID: \"dfd0b89d-949b-459e-ad7c-0a2ef8aa3988\") " pod="service-telemetry/prometheus-webhook-snmp-2-build" Dec 03 00:32:38 crc kubenswrapper[4953]: I1203 00:32:38.400555 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"builder-dockercfg-9prwd-push\" (UniqueName: \"kubernetes.io/secret/dfd0b89d-949b-459e-ad7c-0a2ef8aa3988-builder-dockercfg-9prwd-push\") pod \"prometheus-webhook-snmp-2-build\" (UID: \"dfd0b89d-949b-459e-ad7c-0a2ef8aa3988\") " pod="service-telemetry/prometheus-webhook-snmp-2-build" Dec 03 00:32:38 crc kubenswrapper[4953]: I1203 00:32:38.400801 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/dfd0b89d-949b-459e-ad7c-0a2ef8aa3988-node-pullsecrets\") pod \"prometheus-webhook-snmp-2-build\" (UID: \"dfd0b89d-949b-459e-ad7c-0a2ef8aa3988\") " pod="service-telemetry/prometheus-webhook-snmp-2-build" Dec 03 00:32:38 crc kubenswrapper[4953]: I1203 00:32:38.401300 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/dfd0b89d-949b-459e-ad7c-0a2ef8aa3988-build-blob-cache\") pod \"prometheus-webhook-snmp-2-build\" (UID: \"dfd0b89d-949b-459e-ad7c-0a2ef8aa3988\") " pod="service-telemetry/prometheus-webhook-snmp-2-build" Dec 03 00:32:38 crc kubenswrapper[4953]: I1203 00:32:38.401628 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/dfd0b89d-949b-459e-ad7c-0a2ef8aa3988-build-proxy-ca-bundles\") pod \"prometheus-webhook-snmp-2-build\" (UID: \"dfd0b89d-949b-459e-ad7c-0a2ef8aa3988\") " pod="service-telemetry/prometheus-webhook-snmp-2-build" Dec 03 00:32:38 crc kubenswrapper[4953]: I1203 00:32:38.401634 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/dfd0b89d-949b-459e-ad7c-0a2ef8aa3988-container-storage-root\") pod \"prometheus-webhook-snmp-2-build\" (UID: \"dfd0b89d-949b-459e-ad7c-0a2ef8aa3988\") " pod="service-telemetry/prometheus-webhook-snmp-2-build" Dec 03 00:32:38 crc kubenswrapper[4953]: I1203 00:32:38.401695 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/dfd0b89d-949b-459e-ad7c-0a2ef8aa3988-buildworkdir\") pod \"prometheus-webhook-snmp-2-build\" (UID: \"dfd0b89d-949b-459e-ad7c-0a2ef8aa3988\") " pod="service-telemetry/prometheus-webhook-snmp-2-build" Dec 03 00:32:38 crc kubenswrapper[4953]: I1203 00:32:38.401712 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/dfd0b89d-949b-459e-ad7c-0a2ef8aa3988-build-system-configs\") pod \"prometheus-webhook-snmp-2-build\" (UID: \"dfd0b89d-949b-459e-ad7c-0a2ef8aa3988\") " pod="service-telemetry/prometheus-webhook-snmp-2-build" Dec 03 00:32:38 crc kubenswrapper[4953]: I1203 00:32:38.402003 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/dfd0b89d-949b-459e-ad7c-0a2ef8aa3988-build-ca-bundles\") pod \"prometheus-webhook-snmp-2-build\" (UID: \"dfd0b89d-949b-459e-ad7c-0a2ef8aa3988\") " pod="service-telemetry/prometheus-webhook-snmp-2-build" Dec 03 00:32:38 crc kubenswrapper[4953]: I1203 00:32:38.402026 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/dfd0b89d-949b-459e-ad7c-0a2ef8aa3988-container-storage-run\") pod \"prometheus-webhook-snmp-2-build\" (UID: \"dfd0b89d-949b-459e-ad7c-0a2ef8aa3988\") " pod="service-telemetry/prometheus-webhook-snmp-2-build" Dec 03 00:32:38 crc kubenswrapper[4953]: I1203 00:32:38.406251 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"builder-dockercfg-9prwd-push\" (UniqueName: \"kubernetes.io/secret/dfd0b89d-949b-459e-ad7c-0a2ef8aa3988-builder-dockercfg-9prwd-push\") pod \"prometheus-webhook-snmp-2-build\" (UID: \"dfd0b89d-949b-459e-ad7c-0a2ef8aa3988\") " pod="service-telemetry/prometheus-webhook-snmp-2-build" Dec 03 00:32:38 crc kubenswrapper[4953]: I1203 00:32:38.406320 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"builder-dockercfg-9prwd-pull\" (UniqueName: \"kubernetes.io/secret/dfd0b89d-949b-459e-ad7c-0a2ef8aa3988-builder-dockercfg-9prwd-pull\") pod \"prometheus-webhook-snmp-2-build\" (UID: \"dfd0b89d-949b-459e-ad7c-0a2ef8aa3988\") " pod="service-telemetry/prometheus-webhook-snmp-2-build" Dec 03 00:32:38 crc kubenswrapper[4953]: I1203 00:32:38.431420 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9rlpp\" (UniqueName: \"kubernetes.io/projected/dfd0b89d-949b-459e-ad7c-0a2ef8aa3988-kube-api-access-9rlpp\") pod \"prometheus-webhook-snmp-2-build\" (UID: \"dfd0b89d-949b-459e-ad7c-0a2ef8aa3988\") " pod="service-telemetry/prometheus-webhook-snmp-2-build" Dec 03 00:32:38 crc kubenswrapper[4953]: I1203 00:32:38.457125 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/prometheus-webhook-snmp-2-build" Dec 03 00:32:38 crc kubenswrapper[4953]: I1203 00:32:38.689725 4953 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/prometheus-webhook-snmp-2-build"] Dec 03 00:32:38 crc kubenswrapper[4953]: I1203 00:32:38.739313 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/prometheus-webhook-snmp-2-build" event={"ID":"dfd0b89d-949b-459e-ad7c-0a2ef8aa3988","Type":"ContainerStarted","Data":"efd0443d2a46b9460826367d0f4fac971d6c854dbf6290f398f5a706fbd1b320"} Dec 03 00:32:39 crc kubenswrapper[4953]: I1203 00:32:39.357438 4953 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5dbca6cf-05f2-487a-9d13-c5b9cc9c8740" path="/var/lib/kubelet/pods/5dbca6cf-05f2-487a-9d13-c5b9cc9c8740/volumes" Dec 03 00:32:39 crc kubenswrapper[4953]: I1203 00:32:39.749446 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/prometheus-webhook-snmp-2-build" event={"ID":"dfd0b89d-949b-459e-ad7c-0a2ef8aa3988","Type":"ContainerStarted","Data":"5ce95d026c8de4ada333ffdf1515bfc6d7518bafa002a5d3f9cf3ae4ed622cea"} Dec 03 00:32:39 crc kubenswrapper[4953]: E1203 00:32:39.878520 4953 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.102.83.158:39106->38.102.83.158:33583: write tcp 38.102.83.158:39106->38.102.83.158:33583: write: broken pipe Dec 03 00:32:40 crc kubenswrapper[4953]: I1203 00:32:40.760632 4953 generic.go:334] "Generic (PLEG): container finished" podID="dfd0b89d-949b-459e-ad7c-0a2ef8aa3988" containerID="5ce95d026c8de4ada333ffdf1515bfc6d7518bafa002a5d3f9cf3ae4ed622cea" exitCode=0 Dec 03 00:32:40 crc kubenswrapper[4953]: I1203 00:32:40.760704 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/prometheus-webhook-snmp-2-build" event={"ID":"dfd0b89d-949b-459e-ad7c-0a2ef8aa3988","Type":"ContainerDied","Data":"5ce95d026c8de4ada333ffdf1515bfc6d7518bafa002a5d3f9cf3ae4ed622cea"} Dec 03 00:32:41 crc kubenswrapper[4953]: I1203 00:32:41.772187 4953 generic.go:334] "Generic (PLEG): container finished" podID="dfd0b89d-949b-459e-ad7c-0a2ef8aa3988" containerID="03c193c42fe65a765602e8b2b2c1b1139c771f8550bb70fab60c1b03ad5c440b" exitCode=0 Dec 03 00:32:41 crc kubenswrapper[4953]: I1203 00:32:41.772272 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/prometheus-webhook-snmp-2-build" event={"ID":"dfd0b89d-949b-459e-ad7c-0a2ef8aa3988","Type":"ContainerDied","Data":"03c193c42fe65a765602e8b2b2c1b1139c771f8550bb70fab60c1b03ad5c440b"} Dec 03 00:32:41 crc kubenswrapper[4953]: I1203 00:32:41.817101 4953 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_prometheus-webhook-snmp-2-build_dfd0b89d-949b-459e-ad7c-0a2ef8aa3988/manage-dockerfile/0.log" Dec 03 00:32:42 crc kubenswrapper[4953]: I1203 00:32:42.789602 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/prometheus-webhook-snmp-2-build" event={"ID":"dfd0b89d-949b-459e-ad7c-0a2ef8aa3988","Type":"ContainerStarted","Data":"f14061919feb48aee64548873f7fe581966af0163363891e6d94d85ece2ecbe0"} Dec 03 00:32:42 crc kubenswrapper[4953]: I1203 00:32:42.831614 4953 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="service-telemetry/prometheus-webhook-snmp-2-build" podStartSLOduration=4.831590682 podStartE2EDuration="4.831590682s" podCreationTimestamp="2025-12-03 00:32:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 00:32:42.82701156 +0000 UTC m=+1551.811292343" watchObservedRunningTime="2025-12-03 00:32:42.831590682 +0000 UTC m=+1551.815871465" Dec 03 00:33:18 crc kubenswrapper[4953]: I1203 00:33:18.944234 4953 patch_prober.go:28] interesting pod/machine-config-daemon-6dfbn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 00:33:18 crc kubenswrapper[4953]: I1203 00:33:18.945156 4953 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-6dfbn" podUID="53ffb1e9-9a2e-4869-80d7-776f4aa7ef8b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 00:33:42 crc kubenswrapper[4953]: I1203 00:33:42.280714 4953 generic.go:334] "Generic (PLEG): container finished" podID="dfd0b89d-949b-459e-ad7c-0a2ef8aa3988" containerID="f14061919feb48aee64548873f7fe581966af0163363891e6d94d85ece2ecbe0" exitCode=0 Dec 03 00:33:42 crc kubenswrapper[4953]: I1203 00:33:42.280934 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/prometheus-webhook-snmp-2-build" event={"ID":"dfd0b89d-949b-459e-ad7c-0a2ef8aa3988","Type":"ContainerDied","Data":"f14061919feb48aee64548873f7fe581966af0163363891e6d94d85ece2ecbe0"} Dec 03 00:33:43 crc kubenswrapper[4953]: I1203 00:33:43.549662 4953 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/prometheus-webhook-snmp-2-build" Dec 03 00:33:43 crc kubenswrapper[4953]: I1203 00:33:43.725209 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"builder-dockercfg-9prwd-pull\" (UniqueName: \"kubernetes.io/secret/dfd0b89d-949b-459e-ad7c-0a2ef8aa3988-builder-dockercfg-9prwd-pull\") pod \"dfd0b89d-949b-459e-ad7c-0a2ef8aa3988\" (UID: \"dfd0b89d-949b-459e-ad7c-0a2ef8aa3988\") " Dec 03 00:33:43 crc kubenswrapper[4953]: I1203 00:33:43.725387 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/dfd0b89d-949b-459e-ad7c-0a2ef8aa3988-buildcachedir\") pod \"dfd0b89d-949b-459e-ad7c-0a2ef8aa3988\" (UID: \"dfd0b89d-949b-459e-ad7c-0a2ef8aa3988\") " Dec 03 00:33:43 crc kubenswrapper[4953]: I1203 00:33:43.725500 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/dfd0b89d-949b-459e-ad7c-0a2ef8aa3988-buildcachedir" (OuterVolumeSpecName: "buildcachedir") pod "dfd0b89d-949b-459e-ad7c-0a2ef8aa3988" (UID: "dfd0b89d-949b-459e-ad7c-0a2ef8aa3988"). InnerVolumeSpecName "buildcachedir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 00:33:43 crc kubenswrapper[4953]: I1203 00:33:43.725526 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/dfd0b89d-949b-459e-ad7c-0a2ef8aa3988-build-proxy-ca-bundles\") pod \"dfd0b89d-949b-459e-ad7c-0a2ef8aa3988\" (UID: \"dfd0b89d-949b-459e-ad7c-0a2ef8aa3988\") " Dec 03 00:33:43 crc kubenswrapper[4953]: I1203 00:33:43.725670 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/dfd0b89d-949b-459e-ad7c-0a2ef8aa3988-build-blob-cache\") pod \"dfd0b89d-949b-459e-ad7c-0a2ef8aa3988\" (UID: \"dfd0b89d-949b-459e-ad7c-0a2ef8aa3988\") " Dec 03 00:33:43 crc kubenswrapper[4953]: I1203 00:33:43.725689 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/dfd0b89d-949b-459e-ad7c-0a2ef8aa3988-node-pullsecrets\") pod \"dfd0b89d-949b-459e-ad7c-0a2ef8aa3988\" (UID: \"dfd0b89d-949b-459e-ad7c-0a2ef8aa3988\") " Dec 03 00:33:43 crc kubenswrapper[4953]: I1203 00:33:43.725850 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/dfd0b89d-949b-459e-ad7c-0a2ef8aa3988-node-pullsecrets" (OuterVolumeSpecName: "node-pullsecrets") pod "dfd0b89d-949b-459e-ad7c-0a2ef8aa3988" (UID: "dfd0b89d-949b-459e-ad7c-0a2ef8aa3988"). InnerVolumeSpecName "node-pullsecrets". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 00:33:43 crc kubenswrapper[4953]: I1203 00:33:43.726275 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dfd0b89d-949b-459e-ad7c-0a2ef8aa3988-build-proxy-ca-bundles" (OuterVolumeSpecName: "build-proxy-ca-bundles") pod "dfd0b89d-949b-459e-ad7c-0a2ef8aa3988" (UID: "dfd0b89d-949b-459e-ad7c-0a2ef8aa3988"). InnerVolumeSpecName "build-proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:33:43 crc kubenswrapper[4953]: I1203 00:33:43.727051 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/dfd0b89d-949b-459e-ad7c-0a2ef8aa3988-build-ca-bundles\") pod \"dfd0b89d-949b-459e-ad7c-0a2ef8aa3988\" (UID: \"dfd0b89d-949b-459e-ad7c-0a2ef8aa3988\") " Dec 03 00:33:43 crc kubenswrapper[4953]: I1203 00:33:43.727131 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/dfd0b89d-949b-459e-ad7c-0a2ef8aa3988-buildworkdir\") pod \"dfd0b89d-949b-459e-ad7c-0a2ef8aa3988\" (UID: \"dfd0b89d-949b-459e-ad7c-0a2ef8aa3988\") " Dec 03 00:33:43 crc kubenswrapper[4953]: I1203 00:33:43.727171 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9rlpp\" (UniqueName: \"kubernetes.io/projected/dfd0b89d-949b-459e-ad7c-0a2ef8aa3988-kube-api-access-9rlpp\") pod \"dfd0b89d-949b-459e-ad7c-0a2ef8aa3988\" (UID: \"dfd0b89d-949b-459e-ad7c-0a2ef8aa3988\") " Dec 03 00:33:43 crc kubenswrapper[4953]: I1203 00:33:43.727155 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dfd0b89d-949b-459e-ad7c-0a2ef8aa3988-build-ca-bundles" (OuterVolumeSpecName: "build-ca-bundles") pod "dfd0b89d-949b-459e-ad7c-0a2ef8aa3988" (UID: "dfd0b89d-949b-459e-ad7c-0a2ef8aa3988"). InnerVolumeSpecName "build-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:33:43 crc kubenswrapper[4953]: I1203 00:33:43.727209 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/dfd0b89d-949b-459e-ad7c-0a2ef8aa3988-container-storage-root\") pod \"dfd0b89d-949b-459e-ad7c-0a2ef8aa3988\" (UID: \"dfd0b89d-949b-459e-ad7c-0a2ef8aa3988\") " Dec 03 00:33:43 crc kubenswrapper[4953]: I1203 00:33:43.727233 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/dfd0b89d-949b-459e-ad7c-0a2ef8aa3988-build-system-configs\") pod \"dfd0b89d-949b-459e-ad7c-0a2ef8aa3988\" (UID: \"dfd0b89d-949b-459e-ad7c-0a2ef8aa3988\") " Dec 03 00:33:43 crc kubenswrapper[4953]: I1203 00:33:43.727258 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/dfd0b89d-949b-459e-ad7c-0a2ef8aa3988-container-storage-run\") pod \"dfd0b89d-949b-459e-ad7c-0a2ef8aa3988\" (UID: \"dfd0b89d-949b-459e-ad7c-0a2ef8aa3988\") " Dec 03 00:33:43 crc kubenswrapper[4953]: I1203 00:33:43.727281 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"builder-dockercfg-9prwd-push\" (UniqueName: \"kubernetes.io/secret/dfd0b89d-949b-459e-ad7c-0a2ef8aa3988-builder-dockercfg-9prwd-push\") pod \"dfd0b89d-949b-459e-ad7c-0a2ef8aa3988\" (UID: \"dfd0b89d-949b-459e-ad7c-0a2ef8aa3988\") " Dec 03 00:33:43 crc kubenswrapper[4953]: I1203 00:33:43.727512 4953 reconciler_common.go:293] "Volume detached for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/dfd0b89d-949b-459e-ad7c-0a2ef8aa3988-buildcachedir\") on node \"crc\" DevicePath \"\"" Dec 03 00:33:43 crc kubenswrapper[4953]: I1203 00:33:43.727531 4953 reconciler_common.go:293] "Volume detached for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/dfd0b89d-949b-459e-ad7c-0a2ef8aa3988-build-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Dec 03 00:33:43 crc kubenswrapper[4953]: I1203 00:33:43.727542 4953 reconciler_common.go:293] "Volume detached for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/dfd0b89d-949b-459e-ad7c-0a2ef8aa3988-node-pullsecrets\") on node \"crc\" DevicePath \"\"" Dec 03 00:33:43 crc kubenswrapper[4953]: I1203 00:33:43.727553 4953 reconciler_common.go:293] "Volume detached for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/dfd0b89d-949b-459e-ad7c-0a2ef8aa3988-build-ca-bundles\") on node \"crc\" DevicePath \"\"" Dec 03 00:33:43 crc kubenswrapper[4953]: I1203 00:33:43.730057 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dfd0b89d-949b-459e-ad7c-0a2ef8aa3988-buildworkdir" (OuterVolumeSpecName: "buildworkdir") pod "dfd0b89d-949b-459e-ad7c-0a2ef8aa3988" (UID: "dfd0b89d-949b-459e-ad7c-0a2ef8aa3988"). InnerVolumeSpecName "buildworkdir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 00:33:43 crc kubenswrapper[4953]: I1203 00:33:43.730966 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dfd0b89d-949b-459e-ad7c-0a2ef8aa3988-container-storage-run" (OuterVolumeSpecName: "container-storage-run") pod "dfd0b89d-949b-459e-ad7c-0a2ef8aa3988" (UID: "dfd0b89d-949b-459e-ad7c-0a2ef8aa3988"). InnerVolumeSpecName "container-storage-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 00:33:43 crc kubenswrapper[4953]: I1203 00:33:43.731745 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dfd0b89d-949b-459e-ad7c-0a2ef8aa3988-build-system-configs" (OuterVolumeSpecName: "build-system-configs") pod "dfd0b89d-949b-459e-ad7c-0a2ef8aa3988" (UID: "dfd0b89d-949b-459e-ad7c-0a2ef8aa3988"). InnerVolumeSpecName "build-system-configs". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:33:43 crc kubenswrapper[4953]: I1203 00:33:43.735083 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dfd0b89d-949b-459e-ad7c-0a2ef8aa3988-builder-dockercfg-9prwd-pull" (OuterVolumeSpecName: "builder-dockercfg-9prwd-pull") pod "dfd0b89d-949b-459e-ad7c-0a2ef8aa3988" (UID: "dfd0b89d-949b-459e-ad7c-0a2ef8aa3988"). InnerVolumeSpecName "builder-dockercfg-9prwd-pull". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:33:43 crc kubenswrapper[4953]: I1203 00:33:43.735672 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dfd0b89d-949b-459e-ad7c-0a2ef8aa3988-kube-api-access-9rlpp" (OuterVolumeSpecName: "kube-api-access-9rlpp") pod "dfd0b89d-949b-459e-ad7c-0a2ef8aa3988" (UID: "dfd0b89d-949b-459e-ad7c-0a2ef8aa3988"). InnerVolumeSpecName "kube-api-access-9rlpp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:33:43 crc kubenswrapper[4953]: I1203 00:33:43.742631 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dfd0b89d-949b-459e-ad7c-0a2ef8aa3988-builder-dockercfg-9prwd-push" (OuterVolumeSpecName: "builder-dockercfg-9prwd-push") pod "dfd0b89d-949b-459e-ad7c-0a2ef8aa3988" (UID: "dfd0b89d-949b-459e-ad7c-0a2ef8aa3988"). InnerVolumeSpecName "builder-dockercfg-9prwd-push". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:33:43 crc kubenswrapper[4953]: I1203 00:33:43.825179 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dfd0b89d-949b-459e-ad7c-0a2ef8aa3988-build-blob-cache" (OuterVolumeSpecName: "build-blob-cache") pod "dfd0b89d-949b-459e-ad7c-0a2ef8aa3988" (UID: "dfd0b89d-949b-459e-ad7c-0a2ef8aa3988"). InnerVolumeSpecName "build-blob-cache". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 00:33:43 crc kubenswrapper[4953]: I1203 00:33:43.828724 4953 reconciler_common.go:293] "Volume detached for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/dfd0b89d-949b-459e-ad7c-0a2ef8aa3988-build-blob-cache\") on node \"crc\" DevicePath \"\"" Dec 03 00:33:43 crc kubenswrapper[4953]: I1203 00:33:43.828759 4953 reconciler_common.go:293] "Volume detached for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/dfd0b89d-949b-459e-ad7c-0a2ef8aa3988-buildworkdir\") on node \"crc\" DevicePath \"\"" Dec 03 00:33:43 crc kubenswrapper[4953]: I1203 00:33:43.828770 4953 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9rlpp\" (UniqueName: \"kubernetes.io/projected/dfd0b89d-949b-459e-ad7c-0a2ef8aa3988-kube-api-access-9rlpp\") on node \"crc\" DevicePath \"\"" Dec 03 00:33:43 crc kubenswrapper[4953]: I1203 00:33:43.828780 4953 reconciler_common.go:293] "Volume detached for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/dfd0b89d-949b-459e-ad7c-0a2ef8aa3988-build-system-configs\") on node \"crc\" DevicePath \"\"" Dec 03 00:33:43 crc kubenswrapper[4953]: I1203 00:33:43.828791 4953 reconciler_common.go:293] "Volume detached for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/dfd0b89d-949b-459e-ad7c-0a2ef8aa3988-container-storage-run\") on node \"crc\" DevicePath \"\"" Dec 03 00:33:43 crc kubenswrapper[4953]: I1203 00:33:43.828800 4953 reconciler_common.go:293] "Volume detached for volume \"builder-dockercfg-9prwd-push\" (UniqueName: \"kubernetes.io/secret/dfd0b89d-949b-459e-ad7c-0a2ef8aa3988-builder-dockercfg-9prwd-push\") on node \"crc\" DevicePath \"\"" Dec 03 00:33:43 crc kubenswrapper[4953]: I1203 00:33:43.828813 4953 reconciler_common.go:293] "Volume detached for volume \"builder-dockercfg-9prwd-pull\" (UniqueName: \"kubernetes.io/secret/dfd0b89d-949b-459e-ad7c-0a2ef8aa3988-builder-dockercfg-9prwd-pull\") on node \"crc\" DevicePath \"\"" Dec 03 00:33:44 crc kubenswrapper[4953]: I1203 00:33:44.299255 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/prometheus-webhook-snmp-2-build" event={"ID":"dfd0b89d-949b-459e-ad7c-0a2ef8aa3988","Type":"ContainerDied","Data":"efd0443d2a46b9460826367d0f4fac971d6c854dbf6290f398f5a706fbd1b320"} Dec 03 00:33:44 crc kubenswrapper[4953]: I1203 00:33:44.299305 4953 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="efd0443d2a46b9460826367d0f4fac971d6c854dbf6290f398f5a706fbd1b320" Dec 03 00:33:44 crc kubenswrapper[4953]: I1203 00:33:44.299391 4953 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/prometheus-webhook-snmp-2-build" Dec 03 00:33:44 crc kubenswrapper[4953]: I1203 00:33:44.660733 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dfd0b89d-949b-459e-ad7c-0a2ef8aa3988-container-storage-root" (OuterVolumeSpecName: "container-storage-root") pod "dfd0b89d-949b-459e-ad7c-0a2ef8aa3988" (UID: "dfd0b89d-949b-459e-ad7c-0a2ef8aa3988"). InnerVolumeSpecName "container-storage-root". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 00:33:44 crc kubenswrapper[4953]: I1203 00:33:44.744289 4953 reconciler_common.go:293] "Volume detached for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/dfd0b89d-949b-459e-ad7c-0a2ef8aa3988-container-storage-root\") on node \"crc\" DevicePath \"\"" Dec 03 00:33:48 crc kubenswrapper[4953]: I1203 00:33:48.945097 4953 patch_prober.go:28] interesting pod/machine-config-daemon-6dfbn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 00:33:48 crc kubenswrapper[4953]: I1203 00:33:48.945900 4953 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-6dfbn" podUID="53ffb1e9-9a2e-4869-80d7-776f4aa7ef8b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 00:33:53 crc kubenswrapper[4953]: I1203 00:33:53.449776 4953 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["service-telemetry/service-telemetry-operator-bundle-1-build"] Dec 03 00:33:53 crc kubenswrapper[4953]: E1203 00:33:53.450366 4953 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dfd0b89d-949b-459e-ad7c-0a2ef8aa3988" containerName="manage-dockerfile" Dec 03 00:33:53 crc kubenswrapper[4953]: I1203 00:33:53.450378 4953 state_mem.go:107] "Deleted CPUSet assignment" podUID="dfd0b89d-949b-459e-ad7c-0a2ef8aa3988" containerName="manage-dockerfile" Dec 03 00:33:53 crc kubenswrapper[4953]: E1203 00:33:53.450389 4953 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dfd0b89d-949b-459e-ad7c-0a2ef8aa3988" containerName="git-clone" Dec 03 00:33:53 crc kubenswrapper[4953]: I1203 00:33:53.450396 4953 state_mem.go:107] "Deleted CPUSet assignment" podUID="dfd0b89d-949b-459e-ad7c-0a2ef8aa3988" containerName="git-clone" Dec 03 00:33:53 crc kubenswrapper[4953]: E1203 00:33:53.450410 4953 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dfd0b89d-949b-459e-ad7c-0a2ef8aa3988" containerName="docker-build" Dec 03 00:33:53 crc kubenswrapper[4953]: I1203 00:33:53.450419 4953 state_mem.go:107] "Deleted CPUSet assignment" podUID="dfd0b89d-949b-459e-ad7c-0a2ef8aa3988" containerName="docker-build" Dec 03 00:33:53 crc kubenswrapper[4953]: I1203 00:33:53.450533 4953 memory_manager.go:354] "RemoveStaleState removing state" podUID="dfd0b89d-949b-459e-ad7c-0a2ef8aa3988" containerName="docker-build" Dec 03 00:33:53 crc kubenswrapper[4953]: I1203 00:33:53.451156 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/service-telemetry-operator-bundle-1-build" Dec 03 00:33:53 crc kubenswrapper[4953]: I1203 00:33:53.454196 4953 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"builder-dockercfg-9prwd" Dec 03 00:33:53 crc kubenswrapper[4953]: I1203 00:33:53.454278 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"service-telemetry-operator-bundle-1-ca" Dec 03 00:33:53 crc kubenswrapper[4953]: I1203 00:33:53.455899 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"service-telemetry-operator-bundle-1-sys-config" Dec 03 00:33:53 crc kubenswrapper[4953]: I1203 00:33:53.456086 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"service-telemetry-operator-bundle-1-global-ca" Dec 03 00:33:53 crc kubenswrapper[4953]: I1203 00:33:53.505381 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/9045afb8-248a-4757-bb6c-b6d34c410159-build-proxy-ca-bundles\") pod \"service-telemetry-operator-bundle-1-build\" (UID: \"9045afb8-248a-4757-bb6c-b6d34c410159\") " pod="service-telemetry/service-telemetry-operator-bundle-1-build" Dec 03 00:33:53 crc kubenswrapper[4953]: I1203 00:33:53.505462 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/9045afb8-248a-4757-bb6c-b6d34c410159-build-ca-bundles\") pod \"service-telemetry-operator-bundle-1-build\" (UID: \"9045afb8-248a-4757-bb6c-b6d34c410159\") " pod="service-telemetry/service-telemetry-operator-bundle-1-build" Dec 03 00:33:53 crc kubenswrapper[4953]: I1203 00:33:53.505490 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"builder-dockercfg-9prwd-pull\" (UniqueName: \"kubernetes.io/secret/9045afb8-248a-4757-bb6c-b6d34c410159-builder-dockercfg-9prwd-pull\") pod \"service-telemetry-operator-bundle-1-build\" (UID: \"9045afb8-248a-4757-bb6c-b6d34c410159\") " pod="service-telemetry/service-telemetry-operator-bundle-1-build" Dec 03 00:33:53 crc kubenswrapper[4953]: I1203 00:33:53.505551 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/9045afb8-248a-4757-bb6c-b6d34c410159-build-blob-cache\") pod \"service-telemetry-operator-bundle-1-build\" (UID: \"9045afb8-248a-4757-bb6c-b6d34c410159\") " pod="service-telemetry/service-telemetry-operator-bundle-1-build" Dec 03 00:33:53 crc kubenswrapper[4953]: I1203 00:33:53.505570 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"builder-dockercfg-9prwd-push\" (UniqueName: \"kubernetes.io/secret/9045afb8-248a-4757-bb6c-b6d34c410159-builder-dockercfg-9prwd-push\") pod \"service-telemetry-operator-bundle-1-build\" (UID: \"9045afb8-248a-4757-bb6c-b6d34c410159\") " pod="service-telemetry/service-telemetry-operator-bundle-1-build" Dec 03 00:33:53 crc kubenswrapper[4953]: I1203 00:33:53.505678 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/9045afb8-248a-4757-bb6c-b6d34c410159-container-storage-root\") pod \"service-telemetry-operator-bundle-1-build\" (UID: \"9045afb8-248a-4757-bb6c-b6d34c410159\") " pod="service-telemetry/service-telemetry-operator-bundle-1-build" Dec 03 00:33:53 crc kubenswrapper[4953]: I1203 00:33:53.505731 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/9045afb8-248a-4757-bb6c-b6d34c410159-container-storage-run\") pod \"service-telemetry-operator-bundle-1-build\" (UID: \"9045afb8-248a-4757-bb6c-b6d34c410159\") " pod="service-telemetry/service-telemetry-operator-bundle-1-build" Dec 03 00:33:53 crc kubenswrapper[4953]: I1203 00:33:53.505765 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kmh4n\" (UniqueName: \"kubernetes.io/projected/9045afb8-248a-4757-bb6c-b6d34c410159-kube-api-access-kmh4n\") pod \"service-telemetry-operator-bundle-1-build\" (UID: \"9045afb8-248a-4757-bb6c-b6d34c410159\") " pod="service-telemetry/service-telemetry-operator-bundle-1-build" Dec 03 00:33:53 crc kubenswrapper[4953]: I1203 00:33:53.505830 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/9045afb8-248a-4757-bb6c-b6d34c410159-buildcachedir\") pod \"service-telemetry-operator-bundle-1-build\" (UID: \"9045afb8-248a-4757-bb6c-b6d34c410159\") " pod="service-telemetry/service-telemetry-operator-bundle-1-build" Dec 03 00:33:53 crc kubenswrapper[4953]: I1203 00:33:53.505863 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/9045afb8-248a-4757-bb6c-b6d34c410159-build-system-configs\") pod \"service-telemetry-operator-bundle-1-build\" (UID: \"9045afb8-248a-4757-bb6c-b6d34c410159\") " pod="service-telemetry/service-telemetry-operator-bundle-1-build" Dec 03 00:33:53 crc kubenswrapper[4953]: I1203 00:33:53.505883 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/9045afb8-248a-4757-bb6c-b6d34c410159-node-pullsecrets\") pod \"service-telemetry-operator-bundle-1-build\" (UID: \"9045afb8-248a-4757-bb6c-b6d34c410159\") " pod="service-telemetry/service-telemetry-operator-bundle-1-build" Dec 03 00:33:53 crc kubenswrapper[4953]: I1203 00:33:53.505935 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/9045afb8-248a-4757-bb6c-b6d34c410159-buildworkdir\") pod \"service-telemetry-operator-bundle-1-build\" (UID: \"9045afb8-248a-4757-bb6c-b6d34c410159\") " pod="service-telemetry/service-telemetry-operator-bundle-1-build" Dec 03 00:33:53 crc kubenswrapper[4953]: I1203 00:33:53.519729 4953 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/service-telemetry-operator-bundle-1-build"] Dec 03 00:33:53 crc kubenswrapper[4953]: I1203 00:33:53.607581 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/9045afb8-248a-4757-bb6c-b6d34c410159-container-storage-root\") pod \"service-telemetry-operator-bundle-1-build\" (UID: \"9045afb8-248a-4757-bb6c-b6d34c410159\") " pod="service-telemetry/service-telemetry-operator-bundle-1-build" Dec 03 00:33:53 crc kubenswrapper[4953]: I1203 00:33:53.607652 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/9045afb8-248a-4757-bb6c-b6d34c410159-container-storage-run\") pod \"service-telemetry-operator-bundle-1-build\" (UID: \"9045afb8-248a-4757-bb6c-b6d34c410159\") " pod="service-telemetry/service-telemetry-operator-bundle-1-build" Dec 03 00:33:53 crc kubenswrapper[4953]: I1203 00:33:53.607678 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kmh4n\" (UniqueName: \"kubernetes.io/projected/9045afb8-248a-4757-bb6c-b6d34c410159-kube-api-access-kmh4n\") pod \"service-telemetry-operator-bundle-1-build\" (UID: \"9045afb8-248a-4757-bb6c-b6d34c410159\") " pod="service-telemetry/service-telemetry-operator-bundle-1-build" Dec 03 00:33:53 crc kubenswrapper[4953]: I1203 00:33:53.607705 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/9045afb8-248a-4757-bb6c-b6d34c410159-buildcachedir\") pod \"service-telemetry-operator-bundle-1-build\" (UID: \"9045afb8-248a-4757-bb6c-b6d34c410159\") " pod="service-telemetry/service-telemetry-operator-bundle-1-build" Dec 03 00:33:53 crc kubenswrapper[4953]: I1203 00:33:53.607740 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/9045afb8-248a-4757-bb6c-b6d34c410159-build-system-configs\") pod \"service-telemetry-operator-bundle-1-build\" (UID: \"9045afb8-248a-4757-bb6c-b6d34c410159\") " pod="service-telemetry/service-telemetry-operator-bundle-1-build" Dec 03 00:33:53 crc kubenswrapper[4953]: I1203 00:33:53.607796 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/9045afb8-248a-4757-bb6c-b6d34c410159-node-pullsecrets\") pod \"service-telemetry-operator-bundle-1-build\" (UID: \"9045afb8-248a-4757-bb6c-b6d34c410159\") " pod="service-telemetry/service-telemetry-operator-bundle-1-build" Dec 03 00:33:53 crc kubenswrapper[4953]: I1203 00:33:53.607831 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/9045afb8-248a-4757-bb6c-b6d34c410159-buildworkdir\") pod \"service-telemetry-operator-bundle-1-build\" (UID: \"9045afb8-248a-4757-bb6c-b6d34c410159\") " pod="service-telemetry/service-telemetry-operator-bundle-1-build" Dec 03 00:33:53 crc kubenswrapper[4953]: I1203 00:33:53.607858 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/9045afb8-248a-4757-bb6c-b6d34c410159-build-proxy-ca-bundles\") pod \"service-telemetry-operator-bundle-1-build\" (UID: \"9045afb8-248a-4757-bb6c-b6d34c410159\") " pod="service-telemetry/service-telemetry-operator-bundle-1-build" Dec 03 00:33:53 crc kubenswrapper[4953]: I1203 00:33:53.607895 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/9045afb8-248a-4757-bb6c-b6d34c410159-build-ca-bundles\") pod \"service-telemetry-operator-bundle-1-build\" (UID: \"9045afb8-248a-4757-bb6c-b6d34c410159\") " pod="service-telemetry/service-telemetry-operator-bundle-1-build" Dec 03 00:33:53 crc kubenswrapper[4953]: I1203 00:33:53.607934 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"builder-dockercfg-9prwd-pull\" (UniqueName: \"kubernetes.io/secret/9045afb8-248a-4757-bb6c-b6d34c410159-builder-dockercfg-9prwd-pull\") pod \"service-telemetry-operator-bundle-1-build\" (UID: \"9045afb8-248a-4757-bb6c-b6d34c410159\") " pod="service-telemetry/service-telemetry-operator-bundle-1-build" Dec 03 00:33:53 crc kubenswrapper[4953]: I1203 00:33:53.607967 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/9045afb8-248a-4757-bb6c-b6d34c410159-build-blob-cache\") pod \"service-telemetry-operator-bundle-1-build\" (UID: \"9045afb8-248a-4757-bb6c-b6d34c410159\") " pod="service-telemetry/service-telemetry-operator-bundle-1-build" Dec 03 00:33:53 crc kubenswrapper[4953]: I1203 00:33:53.608004 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"builder-dockercfg-9prwd-push\" (UniqueName: \"kubernetes.io/secret/9045afb8-248a-4757-bb6c-b6d34c410159-builder-dockercfg-9prwd-push\") pod \"service-telemetry-operator-bundle-1-build\" (UID: \"9045afb8-248a-4757-bb6c-b6d34c410159\") " pod="service-telemetry/service-telemetry-operator-bundle-1-build" Dec 03 00:33:53 crc kubenswrapper[4953]: I1203 00:33:53.608620 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/9045afb8-248a-4757-bb6c-b6d34c410159-buildcachedir\") pod \"service-telemetry-operator-bundle-1-build\" (UID: \"9045afb8-248a-4757-bb6c-b6d34c410159\") " pod="service-telemetry/service-telemetry-operator-bundle-1-build" Dec 03 00:33:53 crc kubenswrapper[4953]: I1203 00:33:53.608882 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/9045afb8-248a-4757-bb6c-b6d34c410159-node-pullsecrets\") pod \"service-telemetry-operator-bundle-1-build\" (UID: \"9045afb8-248a-4757-bb6c-b6d34c410159\") " pod="service-telemetry/service-telemetry-operator-bundle-1-build" Dec 03 00:33:53 crc kubenswrapper[4953]: I1203 00:33:53.608987 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/9045afb8-248a-4757-bb6c-b6d34c410159-container-storage-root\") pod \"service-telemetry-operator-bundle-1-build\" (UID: \"9045afb8-248a-4757-bb6c-b6d34c410159\") " pod="service-telemetry/service-telemetry-operator-bundle-1-build" Dec 03 00:33:53 crc kubenswrapper[4953]: I1203 00:33:53.609553 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/9045afb8-248a-4757-bb6c-b6d34c410159-container-storage-run\") pod \"service-telemetry-operator-bundle-1-build\" (UID: \"9045afb8-248a-4757-bb6c-b6d34c410159\") " pod="service-telemetry/service-telemetry-operator-bundle-1-build" Dec 03 00:33:53 crc kubenswrapper[4953]: I1203 00:33:53.609614 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/9045afb8-248a-4757-bb6c-b6d34c410159-buildworkdir\") pod \"service-telemetry-operator-bundle-1-build\" (UID: \"9045afb8-248a-4757-bb6c-b6d34c410159\") " pod="service-telemetry/service-telemetry-operator-bundle-1-build" Dec 03 00:33:53 crc kubenswrapper[4953]: I1203 00:33:53.609843 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/9045afb8-248a-4757-bb6c-b6d34c410159-build-blob-cache\") pod \"service-telemetry-operator-bundle-1-build\" (UID: \"9045afb8-248a-4757-bb6c-b6d34c410159\") " pod="service-telemetry/service-telemetry-operator-bundle-1-build" Dec 03 00:33:53 crc kubenswrapper[4953]: I1203 00:33:53.609948 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/9045afb8-248a-4757-bb6c-b6d34c410159-build-proxy-ca-bundles\") pod \"service-telemetry-operator-bundle-1-build\" (UID: \"9045afb8-248a-4757-bb6c-b6d34c410159\") " pod="service-telemetry/service-telemetry-operator-bundle-1-build" Dec 03 00:33:53 crc kubenswrapper[4953]: I1203 00:33:53.610632 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/9045afb8-248a-4757-bb6c-b6d34c410159-build-system-configs\") pod \"service-telemetry-operator-bundle-1-build\" (UID: \"9045afb8-248a-4757-bb6c-b6d34c410159\") " pod="service-telemetry/service-telemetry-operator-bundle-1-build" Dec 03 00:33:53 crc kubenswrapper[4953]: I1203 00:33:53.610872 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/9045afb8-248a-4757-bb6c-b6d34c410159-build-ca-bundles\") pod \"service-telemetry-operator-bundle-1-build\" (UID: \"9045afb8-248a-4757-bb6c-b6d34c410159\") " pod="service-telemetry/service-telemetry-operator-bundle-1-build" Dec 03 00:33:53 crc kubenswrapper[4953]: I1203 00:33:53.617208 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"builder-dockercfg-9prwd-push\" (UniqueName: \"kubernetes.io/secret/9045afb8-248a-4757-bb6c-b6d34c410159-builder-dockercfg-9prwd-push\") pod \"service-telemetry-operator-bundle-1-build\" (UID: \"9045afb8-248a-4757-bb6c-b6d34c410159\") " pod="service-telemetry/service-telemetry-operator-bundle-1-build" Dec 03 00:33:53 crc kubenswrapper[4953]: I1203 00:33:53.618703 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"builder-dockercfg-9prwd-pull\" (UniqueName: \"kubernetes.io/secret/9045afb8-248a-4757-bb6c-b6d34c410159-builder-dockercfg-9prwd-pull\") pod \"service-telemetry-operator-bundle-1-build\" (UID: \"9045afb8-248a-4757-bb6c-b6d34c410159\") " pod="service-telemetry/service-telemetry-operator-bundle-1-build" Dec 03 00:33:53 crc kubenswrapper[4953]: I1203 00:33:53.629495 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kmh4n\" (UniqueName: \"kubernetes.io/projected/9045afb8-248a-4757-bb6c-b6d34c410159-kube-api-access-kmh4n\") pod \"service-telemetry-operator-bundle-1-build\" (UID: \"9045afb8-248a-4757-bb6c-b6d34c410159\") " pod="service-telemetry/service-telemetry-operator-bundle-1-build" Dec 03 00:33:53 crc kubenswrapper[4953]: I1203 00:33:53.771491 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/service-telemetry-operator-bundle-1-build" Dec 03 00:33:54 crc kubenswrapper[4953]: I1203 00:33:54.030705 4953 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/service-telemetry-operator-bundle-1-build"] Dec 03 00:33:54 crc kubenswrapper[4953]: I1203 00:33:54.369771 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/service-telemetry-operator-bundle-1-build" event={"ID":"9045afb8-248a-4757-bb6c-b6d34c410159","Type":"ContainerStarted","Data":"d99c90dd4eaa1c9cb6f801e55792f830216ccc478d40b00ad5df91a8aade3d6c"} Dec 03 00:33:55 crc kubenswrapper[4953]: I1203 00:33:55.380374 4953 generic.go:334] "Generic (PLEG): container finished" podID="9045afb8-248a-4757-bb6c-b6d34c410159" containerID="2dc6fee63c189a01beee6babd1ef39aac49ee77e0084c0f2e4a36d2bc9f18d9c" exitCode=0 Dec 03 00:33:55 crc kubenswrapper[4953]: I1203 00:33:55.380479 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/service-telemetry-operator-bundle-1-build" event={"ID":"9045afb8-248a-4757-bb6c-b6d34c410159","Type":"ContainerDied","Data":"2dc6fee63c189a01beee6babd1ef39aac49ee77e0084c0f2e4a36d2bc9f18d9c"} Dec 03 00:33:56 crc kubenswrapper[4953]: I1203 00:33:56.389527 4953 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_service-telemetry-operator-bundle-1-build_9045afb8-248a-4757-bb6c-b6d34c410159/docker-build/0.log" Dec 03 00:33:56 crc kubenswrapper[4953]: I1203 00:33:56.390701 4953 generic.go:334] "Generic (PLEG): container finished" podID="9045afb8-248a-4757-bb6c-b6d34c410159" containerID="c3f17a1d41fe0ce559e682dccbc4fcd8027a2761d9c8f4187ff4eb27a32e90b7" exitCode=1 Dec 03 00:33:56 crc kubenswrapper[4953]: I1203 00:33:56.390778 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/service-telemetry-operator-bundle-1-build" event={"ID":"9045afb8-248a-4757-bb6c-b6d34c410159","Type":"ContainerDied","Data":"c3f17a1d41fe0ce559e682dccbc4fcd8027a2761d9c8f4187ff4eb27a32e90b7"} Dec 03 00:33:57 crc kubenswrapper[4953]: I1203 00:33:57.773314 4953 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_service-telemetry-operator-bundle-1-build_9045afb8-248a-4757-bb6c-b6d34c410159/docker-build/0.log" Dec 03 00:33:57 crc kubenswrapper[4953]: I1203 00:33:57.774024 4953 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/service-telemetry-operator-bundle-1-build" Dec 03 00:33:57 crc kubenswrapper[4953]: I1203 00:33:57.883843 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/9045afb8-248a-4757-bb6c-b6d34c410159-container-storage-root\") pod \"9045afb8-248a-4757-bb6c-b6d34c410159\" (UID: \"9045afb8-248a-4757-bb6c-b6d34c410159\") " Dec 03 00:33:57 crc kubenswrapper[4953]: I1203 00:33:57.883889 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/9045afb8-248a-4757-bb6c-b6d34c410159-buildworkdir\") pod \"9045afb8-248a-4757-bb6c-b6d34c410159\" (UID: \"9045afb8-248a-4757-bb6c-b6d34c410159\") " Dec 03 00:33:57 crc kubenswrapper[4953]: I1203 00:33:57.883934 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/9045afb8-248a-4757-bb6c-b6d34c410159-build-proxy-ca-bundles\") pod \"9045afb8-248a-4757-bb6c-b6d34c410159\" (UID: \"9045afb8-248a-4757-bb6c-b6d34c410159\") " Dec 03 00:33:57 crc kubenswrapper[4953]: I1203 00:33:57.884005 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/9045afb8-248a-4757-bb6c-b6d34c410159-buildcachedir\") pod \"9045afb8-248a-4757-bb6c-b6d34c410159\" (UID: \"9045afb8-248a-4757-bb6c-b6d34c410159\") " Dec 03 00:33:57 crc kubenswrapper[4953]: I1203 00:33:57.884024 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/9045afb8-248a-4757-bb6c-b6d34c410159-build-system-configs\") pod \"9045afb8-248a-4757-bb6c-b6d34c410159\" (UID: \"9045afb8-248a-4757-bb6c-b6d34c410159\") " Dec 03 00:33:57 crc kubenswrapper[4953]: I1203 00:33:57.884043 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/9045afb8-248a-4757-bb6c-b6d34c410159-node-pullsecrets\") pod \"9045afb8-248a-4757-bb6c-b6d34c410159\" (UID: \"9045afb8-248a-4757-bb6c-b6d34c410159\") " Dec 03 00:33:57 crc kubenswrapper[4953]: I1203 00:33:57.884062 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"builder-dockercfg-9prwd-pull\" (UniqueName: \"kubernetes.io/secret/9045afb8-248a-4757-bb6c-b6d34c410159-builder-dockercfg-9prwd-pull\") pod \"9045afb8-248a-4757-bb6c-b6d34c410159\" (UID: \"9045afb8-248a-4757-bb6c-b6d34c410159\") " Dec 03 00:33:57 crc kubenswrapper[4953]: I1203 00:33:57.884092 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kmh4n\" (UniqueName: \"kubernetes.io/projected/9045afb8-248a-4757-bb6c-b6d34c410159-kube-api-access-kmh4n\") pod \"9045afb8-248a-4757-bb6c-b6d34c410159\" (UID: \"9045afb8-248a-4757-bb6c-b6d34c410159\") " Dec 03 00:33:57 crc kubenswrapper[4953]: I1203 00:33:57.884123 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"builder-dockercfg-9prwd-push\" (UniqueName: \"kubernetes.io/secret/9045afb8-248a-4757-bb6c-b6d34c410159-builder-dockercfg-9prwd-push\") pod \"9045afb8-248a-4757-bb6c-b6d34c410159\" (UID: \"9045afb8-248a-4757-bb6c-b6d34c410159\") " Dec 03 00:33:57 crc kubenswrapper[4953]: I1203 00:33:57.884156 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/9045afb8-248a-4757-bb6c-b6d34c410159-build-blob-cache\") pod \"9045afb8-248a-4757-bb6c-b6d34c410159\" (UID: \"9045afb8-248a-4757-bb6c-b6d34c410159\") " Dec 03 00:33:57 crc kubenswrapper[4953]: I1203 00:33:57.884172 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9045afb8-248a-4757-bb6c-b6d34c410159-buildcachedir" (OuterVolumeSpecName: "buildcachedir") pod "9045afb8-248a-4757-bb6c-b6d34c410159" (UID: "9045afb8-248a-4757-bb6c-b6d34c410159"). InnerVolumeSpecName "buildcachedir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 00:33:57 crc kubenswrapper[4953]: I1203 00:33:57.884218 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/9045afb8-248a-4757-bb6c-b6d34c410159-container-storage-run\") pod \"9045afb8-248a-4757-bb6c-b6d34c410159\" (UID: \"9045afb8-248a-4757-bb6c-b6d34c410159\") " Dec 03 00:33:57 crc kubenswrapper[4953]: I1203 00:33:57.884482 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/9045afb8-248a-4757-bb6c-b6d34c410159-build-ca-bundles\") pod \"9045afb8-248a-4757-bb6c-b6d34c410159\" (UID: \"9045afb8-248a-4757-bb6c-b6d34c410159\") " Dec 03 00:33:57 crc kubenswrapper[4953]: I1203 00:33:57.884517 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9045afb8-248a-4757-bb6c-b6d34c410159-node-pullsecrets" (OuterVolumeSpecName: "node-pullsecrets") pod "9045afb8-248a-4757-bb6c-b6d34c410159" (UID: "9045afb8-248a-4757-bb6c-b6d34c410159"). InnerVolumeSpecName "node-pullsecrets". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 00:33:57 crc kubenswrapper[4953]: I1203 00:33:57.885508 4953 reconciler_common.go:293] "Volume detached for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/9045afb8-248a-4757-bb6c-b6d34c410159-buildcachedir\") on node \"crc\" DevicePath \"\"" Dec 03 00:33:57 crc kubenswrapper[4953]: I1203 00:33:57.885541 4953 reconciler_common.go:293] "Volume detached for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/9045afb8-248a-4757-bb6c-b6d34c410159-node-pullsecrets\") on node \"crc\" DevicePath \"\"" Dec 03 00:33:57 crc kubenswrapper[4953]: I1203 00:33:57.885589 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9045afb8-248a-4757-bb6c-b6d34c410159-build-ca-bundles" (OuterVolumeSpecName: "build-ca-bundles") pod "9045afb8-248a-4757-bb6c-b6d34c410159" (UID: "9045afb8-248a-4757-bb6c-b6d34c410159"). InnerVolumeSpecName "build-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:33:57 crc kubenswrapper[4953]: I1203 00:33:57.885623 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9045afb8-248a-4757-bb6c-b6d34c410159-buildworkdir" (OuterVolumeSpecName: "buildworkdir") pod "9045afb8-248a-4757-bb6c-b6d34c410159" (UID: "9045afb8-248a-4757-bb6c-b6d34c410159"). InnerVolumeSpecName "buildworkdir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 00:33:57 crc kubenswrapper[4953]: I1203 00:33:57.885701 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9045afb8-248a-4757-bb6c-b6d34c410159-build-blob-cache" (OuterVolumeSpecName: "build-blob-cache") pod "9045afb8-248a-4757-bb6c-b6d34c410159" (UID: "9045afb8-248a-4757-bb6c-b6d34c410159"). InnerVolumeSpecName "build-blob-cache". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 00:33:57 crc kubenswrapper[4953]: I1203 00:33:57.885839 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9045afb8-248a-4757-bb6c-b6d34c410159-build-system-configs" (OuterVolumeSpecName: "build-system-configs") pod "9045afb8-248a-4757-bb6c-b6d34c410159" (UID: "9045afb8-248a-4757-bb6c-b6d34c410159"). InnerVolumeSpecName "build-system-configs". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:33:57 crc kubenswrapper[4953]: I1203 00:33:57.885966 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9045afb8-248a-4757-bb6c-b6d34c410159-build-proxy-ca-bundles" (OuterVolumeSpecName: "build-proxy-ca-bundles") pod "9045afb8-248a-4757-bb6c-b6d34c410159" (UID: "9045afb8-248a-4757-bb6c-b6d34c410159"). InnerVolumeSpecName "build-proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:33:57 crc kubenswrapper[4953]: I1203 00:33:57.886424 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9045afb8-248a-4757-bb6c-b6d34c410159-container-storage-run" (OuterVolumeSpecName: "container-storage-run") pod "9045afb8-248a-4757-bb6c-b6d34c410159" (UID: "9045afb8-248a-4757-bb6c-b6d34c410159"). InnerVolumeSpecName "container-storage-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 00:33:57 crc kubenswrapper[4953]: I1203 00:33:57.888094 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9045afb8-248a-4757-bb6c-b6d34c410159-container-storage-root" (OuterVolumeSpecName: "container-storage-root") pod "9045afb8-248a-4757-bb6c-b6d34c410159" (UID: "9045afb8-248a-4757-bb6c-b6d34c410159"). InnerVolumeSpecName "container-storage-root". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 00:33:57 crc kubenswrapper[4953]: I1203 00:33:57.891633 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9045afb8-248a-4757-bb6c-b6d34c410159-builder-dockercfg-9prwd-push" (OuterVolumeSpecName: "builder-dockercfg-9prwd-push") pod "9045afb8-248a-4757-bb6c-b6d34c410159" (UID: "9045afb8-248a-4757-bb6c-b6d34c410159"). InnerVolumeSpecName "builder-dockercfg-9prwd-push". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:33:57 crc kubenswrapper[4953]: I1203 00:33:57.892212 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9045afb8-248a-4757-bb6c-b6d34c410159-kube-api-access-kmh4n" (OuterVolumeSpecName: "kube-api-access-kmh4n") pod "9045afb8-248a-4757-bb6c-b6d34c410159" (UID: "9045afb8-248a-4757-bb6c-b6d34c410159"). InnerVolumeSpecName "kube-api-access-kmh4n". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:33:57 crc kubenswrapper[4953]: I1203 00:33:57.892421 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9045afb8-248a-4757-bb6c-b6d34c410159-builder-dockercfg-9prwd-pull" (OuterVolumeSpecName: "builder-dockercfg-9prwd-pull") pod "9045afb8-248a-4757-bb6c-b6d34c410159" (UID: "9045afb8-248a-4757-bb6c-b6d34c410159"). InnerVolumeSpecName "builder-dockercfg-9prwd-pull". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:33:57 crc kubenswrapper[4953]: I1203 00:33:57.987335 4953 reconciler_common.go:293] "Volume detached for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/9045afb8-248a-4757-bb6c-b6d34c410159-container-storage-run\") on node \"crc\" DevicePath \"\"" Dec 03 00:33:57 crc kubenswrapper[4953]: I1203 00:33:57.987902 4953 reconciler_common.go:293] "Volume detached for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/9045afb8-248a-4757-bb6c-b6d34c410159-build-ca-bundles\") on node \"crc\" DevicePath \"\"" Dec 03 00:33:57 crc kubenswrapper[4953]: I1203 00:33:57.987912 4953 reconciler_common.go:293] "Volume detached for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/9045afb8-248a-4757-bb6c-b6d34c410159-container-storage-root\") on node \"crc\" DevicePath \"\"" Dec 03 00:33:57 crc kubenswrapper[4953]: I1203 00:33:57.987924 4953 reconciler_common.go:293] "Volume detached for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/9045afb8-248a-4757-bb6c-b6d34c410159-buildworkdir\") on node \"crc\" DevicePath \"\"" Dec 03 00:33:57 crc kubenswrapper[4953]: I1203 00:33:57.987935 4953 reconciler_common.go:293] "Volume detached for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/9045afb8-248a-4757-bb6c-b6d34c410159-build-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Dec 03 00:33:57 crc kubenswrapper[4953]: I1203 00:33:57.987944 4953 reconciler_common.go:293] "Volume detached for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/9045afb8-248a-4757-bb6c-b6d34c410159-build-system-configs\") on node \"crc\" DevicePath \"\"" Dec 03 00:33:57 crc kubenswrapper[4953]: I1203 00:33:57.987952 4953 reconciler_common.go:293] "Volume detached for volume \"builder-dockercfg-9prwd-pull\" (UniqueName: \"kubernetes.io/secret/9045afb8-248a-4757-bb6c-b6d34c410159-builder-dockercfg-9prwd-pull\") on node \"crc\" DevicePath \"\"" Dec 03 00:33:57 crc kubenswrapper[4953]: I1203 00:33:57.987962 4953 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kmh4n\" (UniqueName: \"kubernetes.io/projected/9045afb8-248a-4757-bb6c-b6d34c410159-kube-api-access-kmh4n\") on node \"crc\" DevicePath \"\"" Dec 03 00:33:57 crc kubenswrapper[4953]: I1203 00:33:57.987990 4953 reconciler_common.go:293] "Volume detached for volume \"builder-dockercfg-9prwd-push\" (UniqueName: \"kubernetes.io/secret/9045afb8-248a-4757-bb6c-b6d34c410159-builder-dockercfg-9prwd-push\") on node \"crc\" DevicePath \"\"" Dec 03 00:33:57 crc kubenswrapper[4953]: I1203 00:33:57.988002 4953 reconciler_common.go:293] "Volume detached for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/9045afb8-248a-4757-bb6c-b6d34c410159-build-blob-cache\") on node \"crc\" DevicePath \"\"" Dec 03 00:33:58 crc kubenswrapper[4953]: I1203 00:33:58.408169 4953 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_service-telemetry-operator-bundle-1-build_9045afb8-248a-4757-bb6c-b6d34c410159/docker-build/0.log" Dec 03 00:33:58 crc kubenswrapper[4953]: I1203 00:33:58.408783 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/service-telemetry-operator-bundle-1-build" event={"ID":"9045afb8-248a-4757-bb6c-b6d34c410159","Type":"ContainerDied","Data":"d99c90dd4eaa1c9cb6f801e55792f830216ccc478d40b00ad5df91a8aade3d6c"} Dec 03 00:33:58 crc kubenswrapper[4953]: I1203 00:33:58.408825 4953 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d99c90dd4eaa1c9cb6f801e55792f830216ccc478d40b00ad5df91a8aade3d6c" Dec 03 00:33:58 crc kubenswrapper[4953]: I1203 00:33:58.408853 4953 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/service-telemetry-operator-bundle-1-build" Dec 03 00:34:03 crc kubenswrapper[4953]: I1203 00:34:03.974835 4953 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["service-telemetry/service-telemetry-operator-bundle-1-build"] Dec 03 00:34:03 crc kubenswrapper[4953]: I1203 00:34:03.986425 4953 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["service-telemetry/service-telemetry-operator-bundle-1-build"] Dec 03 00:34:05 crc kubenswrapper[4953]: I1203 00:34:05.358463 4953 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9045afb8-248a-4757-bb6c-b6d34c410159" path="/var/lib/kubelet/pods/9045afb8-248a-4757-bb6c-b6d34c410159/volumes" Dec 03 00:34:05 crc kubenswrapper[4953]: I1203 00:34:05.612067 4953 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["service-telemetry/service-telemetry-operator-bundle-2-build"] Dec 03 00:34:05 crc kubenswrapper[4953]: E1203 00:34:05.612470 4953 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9045afb8-248a-4757-bb6c-b6d34c410159" containerName="manage-dockerfile" Dec 03 00:34:05 crc kubenswrapper[4953]: I1203 00:34:05.612487 4953 state_mem.go:107] "Deleted CPUSet assignment" podUID="9045afb8-248a-4757-bb6c-b6d34c410159" containerName="manage-dockerfile" Dec 03 00:34:05 crc kubenswrapper[4953]: E1203 00:34:05.612501 4953 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9045afb8-248a-4757-bb6c-b6d34c410159" containerName="docker-build" Dec 03 00:34:05 crc kubenswrapper[4953]: I1203 00:34:05.612509 4953 state_mem.go:107] "Deleted CPUSet assignment" podUID="9045afb8-248a-4757-bb6c-b6d34c410159" containerName="docker-build" Dec 03 00:34:05 crc kubenswrapper[4953]: I1203 00:34:05.612669 4953 memory_manager.go:354] "RemoveStaleState removing state" podUID="9045afb8-248a-4757-bb6c-b6d34c410159" containerName="docker-build" Dec 03 00:34:05 crc kubenswrapper[4953]: I1203 00:34:05.613640 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/service-telemetry-operator-bundle-2-build" Dec 03 00:34:05 crc kubenswrapper[4953]: I1203 00:34:05.618072 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"service-telemetry-operator-bundle-2-global-ca" Dec 03 00:34:05 crc kubenswrapper[4953]: I1203 00:34:05.618126 4953 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/service-telemetry-operator-bundle-2-build"] Dec 03 00:34:05 crc kubenswrapper[4953]: I1203 00:34:05.618408 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"service-telemetry-operator-bundle-2-ca" Dec 03 00:34:05 crc kubenswrapper[4953]: I1203 00:34:05.618528 4953 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"builder-dockercfg-9prwd" Dec 03 00:34:05 crc kubenswrapper[4953]: I1203 00:34:05.618857 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"service-telemetry-operator-bundle-2-sys-config" Dec 03 00:34:05 crc kubenswrapper[4953]: I1203 00:34:05.805941 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mbj7m\" (UniqueName: \"kubernetes.io/projected/3fd62fed-256b-4d94-8757-9271988b9119-kube-api-access-mbj7m\") pod \"service-telemetry-operator-bundle-2-build\" (UID: \"3fd62fed-256b-4d94-8757-9271988b9119\") " pod="service-telemetry/service-telemetry-operator-bundle-2-build" Dec 03 00:34:05 crc kubenswrapper[4953]: I1203 00:34:05.806067 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"builder-dockercfg-9prwd-push\" (UniqueName: \"kubernetes.io/secret/3fd62fed-256b-4d94-8757-9271988b9119-builder-dockercfg-9prwd-push\") pod \"service-telemetry-operator-bundle-2-build\" (UID: \"3fd62fed-256b-4d94-8757-9271988b9119\") " pod="service-telemetry/service-telemetry-operator-bundle-2-build" Dec 03 00:34:05 crc kubenswrapper[4953]: I1203 00:34:05.806110 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/3fd62fed-256b-4d94-8757-9271988b9119-build-ca-bundles\") pod \"service-telemetry-operator-bundle-2-build\" (UID: \"3fd62fed-256b-4d94-8757-9271988b9119\") " pod="service-telemetry/service-telemetry-operator-bundle-2-build" Dec 03 00:34:05 crc kubenswrapper[4953]: I1203 00:34:05.806148 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"builder-dockercfg-9prwd-pull\" (UniqueName: \"kubernetes.io/secret/3fd62fed-256b-4d94-8757-9271988b9119-builder-dockercfg-9prwd-pull\") pod \"service-telemetry-operator-bundle-2-build\" (UID: \"3fd62fed-256b-4d94-8757-9271988b9119\") " pod="service-telemetry/service-telemetry-operator-bundle-2-build" Dec 03 00:34:05 crc kubenswrapper[4953]: I1203 00:34:05.806175 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/3fd62fed-256b-4d94-8757-9271988b9119-buildcachedir\") pod \"service-telemetry-operator-bundle-2-build\" (UID: \"3fd62fed-256b-4d94-8757-9271988b9119\") " pod="service-telemetry/service-telemetry-operator-bundle-2-build" Dec 03 00:34:05 crc kubenswrapper[4953]: I1203 00:34:05.806326 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/3fd62fed-256b-4d94-8757-9271988b9119-buildworkdir\") pod \"service-telemetry-operator-bundle-2-build\" (UID: \"3fd62fed-256b-4d94-8757-9271988b9119\") " pod="service-telemetry/service-telemetry-operator-bundle-2-build" Dec 03 00:34:05 crc kubenswrapper[4953]: I1203 00:34:05.806386 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/3fd62fed-256b-4d94-8757-9271988b9119-container-storage-run\") pod \"service-telemetry-operator-bundle-2-build\" (UID: \"3fd62fed-256b-4d94-8757-9271988b9119\") " pod="service-telemetry/service-telemetry-operator-bundle-2-build" Dec 03 00:34:05 crc kubenswrapper[4953]: I1203 00:34:05.806410 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/3fd62fed-256b-4d94-8757-9271988b9119-node-pullsecrets\") pod \"service-telemetry-operator-bundle-2-build\" (UID: \"3fd62fed-256b-4d94-8757-9271988b9119\") " pod="service-telemetry/service-telemetry-operator-bundle-2-build" Dec 03 00:34:05 crc kubenswrapper[4953]: I1203 00:34:05.806429 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/3fd62fed-256b-4d94-8757-9271988b9119-build-blob-cache\") pod \"service-telemetry-operator-bundle-2-build\" (UID: \"3fd62fed-256b-4d94-8757-9271988b9119\") " pod="service-telemetry/service-telemetry-operator-bundle-2-build" Dec 03 00:34:05 crc kubenswrapper[4953]: I1203 00:34:05.806479 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/3fd62fed-256b-4d94-8757-9271988b9119-build-system-configs\") pod \"service-telemetry-operator-bundle-2-build\" (UID: \"3fd62fed-256b-4d94-8757-9271988b9119\") " pod="service-telemetry/service-telemetry-operator-bundle-2-build" Dec 03 00:34:05 crc kubenswrapper[4953]: I1203 00:34:05.806570 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/3fd62fed-256b-4d94-8757-9271988b9119-build-proxy-ca-bundles\") pod \"service-telemetry-operator-bundle-2-build\" (UID: \"3fd62fed-256b-4d94-8757-9271988b9119\") " pod="service-telemetry/service-telemetry-operator-bundle-2-build" Dec 03 00:34:05 crc kubenswrapper[4953]: I1203 00:34:05.806600 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/3fd62fed-256b-4d94-8757-9271988b9119-container-storage-root\") pod \"service-telemetry-operator-bundle-2-build\" (UID: \"3fd62fed-256b-4d94-8757-9271988b9119\") " pod="service-telemetry/service-telemetry-operator-bundle-2-build" Dec 03 00:34:05 crc kubenswrapper[4953]: I1203 00:34:05.907625 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/3fd62fed-256b-4d94-8757-9271988b9119-build-proxy-ca-bundles\") pod \"service-telemetry-operator-bundle-2-build\" (UID: \"3fd62fed-256b-4d94-8757-9271988b9119\") " pod="service-telemetry/service-telemetry-operator-bundle-2-build" Dec 03 00:34:05 crc kubenswrapper[4953]: I1203 00:34:05.907688 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/3fd62fed-256b-4d94-8757-9271988b9119-container-storage-root\") pod \"service-telemetry-operator-bundle-2-build\" (UID: \"3fd62fed-256b-4d94-8757-9271988b9119\") " pod="service-telemetry/service-telemetry-operator-bundle-2-build" Dec 03 00:34:05 crc kubenswrapper[4953]: I1203 00:34:05.907725 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mbj7m\" (UniqueName: \"kubernetes.io/projected/3fd62fed-256b-4d94-8757-9271988b9119-kube-api-access-mbj7m\") pod \"service-telemetry-operator-bundle-2-build\" (UID: \"3fd62fed-256b-4d94-8757-9271988b9119\") " pod="service-telemetry/service-telemetry-operator-bundle-2-build" Dec 03 00:34:05 crc kubenswrapper[4953]: I1203 00:34:05.907752 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"builder-dockercfg-9prwd-push\" (UniqueName: \"kubernetes.io/secret/3fd62fed-256b-4d94-8757-9271988b9119-builder-dockercfg-9prwd-push\") pod \"service-telemetry-operator-bundle-2-build\" (UID: \"3fd62fed-256b-4d94-8757-9271988b9119\") " pod="service-telemetry/service-telemetry-operator-bundle-2-build" Dec 03 00:34:05 crc kubenswrapper[4953]: I1203 00:34:05.907783 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/3fd62fed-256b-4d94-8757-9271988b9119-build-ca-bundles\") pod \"service-telemetry-operator-bundle-2-build\" (UID: \"3fd62fed-256b-4d94-8757-9271988b9119\") " pod="service-telemetry/service-telemetry-operator-bundle-2-build" Dec 03 00:34:05 crc kubenswrapper[4953]: I1203 00:34:05.907815 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"builder-dockercfg-9prwd-pull\" (UniqueName: \"kubernetes.io/secret/3fd62fed-256b-4d94-8757-9271988b9119-builder-dockercfg-9prwd-pull\") pod \"service-telemetry-operator-bundle-2-build\" (UID: \"3fd62fed-256b-4d94-8757-9271988b9119\") " pod="service-telemetry/service-telemetry-operator-bundle-2-build" Dec 03 00:34:05 crc kubenswrapper[4953]: I1203 00:34:05.907841 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/3fd62fed-256b-4d94-8757-9271988b9119-buildcachedir\") pod \"service-telemetry-operator-bundle-2-build\" (UID: \"3fd62fed-256b-4d94-8757-9271988b9119\") " pod="service-telemetry/service-telemetry-operator-bundle-2-build" Dec 03 00:34:05 crc kubenswrapper[4953]: I1203 00:34:05.907882 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/3fd62fed-256b-4d94-8757-9271988b9119-buildworkdir\") pod \"service-telemetry-operator-bundle-2-build\" (UID: \"3fd62fed-256b-4d94-8757-9271988b9119\") " pod="service-telemetry/service-telemetry-operator-bundle-2-build" Dec 03 00:34:05 crc kubenswrapper[4953]: I1203 00:34:05.907909 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/3fd62fed-256b-4d94-8757-9271988b9119-container-storage-run\") pod \"service-telemetry-operator-bundle-2-build\" (UID: \"3fd62fed-256b-4d94-8757-9271988b9119\") " pod="service-telemetry/service-telemetry-operator-bundle-2-build" Dec 03 00:34:05 crc kubenswrapper[4953]: I1203 00:34:05.907939 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/3fd62fed-256b-4d94-8757-9271988b9119-node-pullsecrets\") pod \"service-telemetry-operator-bundle-2-build\" (UID: \"3fd62fed-256b-4d94-8757-9271988b9119\") " pod="service-telemetry/service-telemetry-operator-bundle-2-build" Dec 03 00:34:05 crc kubenswrapper[4953]: I1203 00:34:05.907965 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/3fd62fed-256b-4d94-8757-9271988b9119-build-blob-cache\") pod \"service-telemetry-operator-bundle-2-build\" (UID: \"3fd62fed-256b-4d94-8757-9271988b9119\") " pod="service-telemetry/service-telemetry-operator-bundle-2-build" Dec 03 00:34:05 crc kubenswrapper[4953]: I1203 00:34:05.908016 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/3fd62fed-256b-4d94-8757-9271988b9119-build-system-configs\") pod \"service-telemetry-operator-bundle-2-build\" (UID: \"3fd62fed-256b-4d94-8757-9271988b9119\") " pod="service-telemetry/service-telemetry-operator-bundle-2-build" Dec 03 00:34:05 crc kubenswrapper[4953]: I1203 00:34:05.908225 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/3fd62fed-256b-4d94-8757-9271988b9119-buildcachedir\") pod \"service-telemetry-operator-bundle-2-build\" (UID: \"3fd62fed-256b-4d94-8757-9271988b9119\") " pod="service-telemetry/service-telemetry-operator-bundle-2-build" Dec 03 00:34:05 crc kubenswrapper[4953]: I1203 00:34:05.908222 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/3fd62fed-256b-4d94-8757-9271988b9119-container-storage-root\") pod \"service-telemetry-operator-bundle-2-build\" (UID: \"3fd62fed-256b-4d94-8757-9271988b9119\") " pod="service-telemetry/service-telemetry-operator-bundle-2-build" Dec 03 00:34:05 crc kubenswrapper[4953]: I1203 00:34:05.908265 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/3fd62fed-256b-4d94-8757-9271988b9119-node-pullsecrets\") pod \"service-telemetry-operator-bundle-2-build\" (UID: \"3fd62fed-256b-4d94-8757-9271988b9119\") " pod="service-telemetry/service-telemetry-operator-bundle-2-build" Dec 03 00:34:05 crc kubenswrapper[4953]: I1203 00:34:05.908653 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/3fd62fed-256b-4d94-8757-9271988b9119-container-storage-run\") pod \"service-telemetry-operator-bundle-2-build\" (UID: \"3fd62fed-256b-4d94-8757-9271988b9119\") " pod="service-telemetry/service-telemetry-operator-bundle-2-build" Dec 03 00:34:05 crc kubenswrapper[4953]: I1203 00:34:05.908708 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/3fd62fed-256b-4d94-8757-9271988b9119-buildworkdir\") pod \"service-telemetry-operator-bundle-2-build\" (UID: \"3fd62fed-256b-4d94-8757-9271988b9119\") " pod="service-telemetry/service-telemetry-operator-bundle-2-build" Dec 03 00:34:05 crc kubenswrapper[4953]: I1203 00:34:05.908968 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/3fd62fed-256b-4d94-8757-9271988b9119-build-system-configs\") pod \"service-telemetry-operator-bundle-2-build\" (UID: \"3fd62fed-256b-4d94-8757-9271988b9119\") " pod="service-telemetry/service-telemetry-operator-bundle-2-build" Dec 03 00:34:05 crc kubenswrapper[4953]: I1203 00:34:05.909040 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/3fd62fed-256b-4d94-8757-9271988b9119-build-blob-cache\") pod \"service-telemetry-operator-bundle-2-build\" (UID: \"3fd62fed-256b-4d94-8757-9271988b9119\") " pod="service-telemetry/service-telemetry-operator-bundle-2-build" Dec 03 00:34:05 crc kubenswrapper[4953]: I1203 00:34:05.909188 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/3fd62fed-256b-4d94-8757-9271988b9119-build-ca-bundles\") pod \"service-telemetry-operator-bundle-2-build\" (UID: \"3fd62fed-256b-4d94-8757-9271988b9119\") " pod="service-telemetry/service-telemetry-operator-bundle-2-build" Dec 03 00:34:05 crc kubenswrapper[4953]: I1203 00:34:05.909427 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/3fd62fed-256b-4d94-8757-9271988b9119-build-proxy-ca-bundles\") pod \"service-telemetry-operator-bundle-2-build\" (UID: \"3fd62fed-256b-4d94-8757-9271988b9119\") " pod="service-telemetry/service-telemetry-operator-bundle-2-build" Dec 03 00:34:05 crc kubenswrapper[4953]: I1203 00:34:05.915267 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"builder-dockercfg-9prwd-pull\" (UniqueName: \"kubernetes.io/secret/3fd62fed-256b-4d94-8757-9271988b9119-builder-dockercfg-9prwd-pull\") pod \"service-telemetry-operator-bundle-2-build\" (UID: \"3fd62fed-256b-4d94-8757-9271988b9119\") " pod="service-telemetry/service-telemetry-operator-bundle-2-build" Dec 03 00:34:05 crc kubenswrapper[4953]: I1203 00:34:05.916647 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"builder-dockercfg-9prwd-push\" (UniqueName: \"kubernetes.io/secret/3fd62fed-256b-4d94-8757-9271988b9119-builder-dockercfg-9prwd-push\") pod \"service-telemetry-operator-bundle-2-build\" (UID: \"3fd62fed-256b-4d94-8757-9271988b9119\") " pod="service-telemetry/service-telemetry-operator-bundle-2-build" Dec 03 00:34:05 crc kubenswrapper[4953]: I1203 00:34:05.924164 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mbj7m\" (UniqueName: \"kubernetes.io/projected/3fd62fed-256b-4d94-8757-9271988b9119-kube-api-access-mbj7m\") pod \"service-telemetry-operator-bundle-2-build\" (UID: \"3fd62fed-256b-4d94-8757-9271988b9119\") " pod="service-telemetry/service-telemetry-operator-bundle-2-build" Dec 03 00:34:05 crc kubenswrapper[4953]: I1203 00:34:05.933182 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/service-telemetry-operator-bundle-2-build" Dec 03 00:34:06 crc kubenswrapper[4953]: I1203 00:34:06.181593 4953 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/service-telemetry-operator-bundle-2-build"] Dec 03 00:34:06 crc kubenswrapper[4953]: I1203 00:34:06.464786 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/service-telemetry-operator-bundle-2-build" event={"ID":"3fd62fed-256b-4d94-8757-9271988b9119","Type":"ContainerStarted","Data":"56aa65cc8e5b3f8376d5dd43fa360a6fb8221d2ca626286ecb48e1009c913592"} Dec 03 00:34:07 crc kubenswrapper[4953]: I1203 00:34:07.473532 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/service-telemetry-operator-bundle-2-build" event={"ID":"3fd62fed-256b-4d94-8757-9271988b9119","Type":"ContainerStarted","Data":"3c9eb25ef22ccc9ee51293db5544305b004081cc21a3028138e9348b7d59550a"} Dec 03 00:34:08 crc kubenswrapper[4953]: I1203 00:34:08.484229 4953 generic.go:334] "Generic (PLEG): container finished" podID="3fd62fed-256b-4d94-8757-9271988b9119" containerID="3c9eb25ef22ccc9ee51293db5544305b004081cc21a3028138e9348b7d59550a" exitCode=0 Dec 03 00:34:08 crc kubenswrapper[4953]: I1203 00:34:08.484373 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/service-telemetry-operator-bundle-2-build" event={"ID":"3fd62fed-256b-4d94-8757-9271988b9119","Type":"ContainerDied","Data":"3c9eb25ef22ccc9ee51293db5544305b004081cc21a3028138e9348b7d59550a"} Dec 03 00:34:09 crc kubenswrapper[4953]: I1203 00:34:09.495183 4953 generic.go:334] "Generic (PLEG): container finished" podID="3fd62fed-256b-4d94-8757-9271988b9119" containerID="b34db421c655e1a6b209fff91244ad09575d6aedc35adb5249da61016e66f1f4" exitCode=0 Dec 03 00:34:09 crc kubenswrapper[4953]: I1203 00:34:09.495242 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/service-telemetry-operator-bundle-2-build" event={"ID":"3fd62fed-256b-4d94-8757-9271988b9119","Type":"ContainerDied","Data":"b34db421c655e1a6b209fff91244ad09575d6aedc35adb5249da61016e66f1f4"} Dec 03 00:34:09 crc kubenswrapper[4953]: I1203 00:34:09.564185 4953 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_service-telemetry-operator-bundle-2-build_3fd62fed-256b-4d94-8757-9271988b9119/manage-dockerfile/0.log" Dec 03 00:34:10 crc kubenswrapper[4953]: I1203 00:34:10.303732 4953 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-5qxfs"] Dec 03 00:34:10 crc kubenswrapper[4953]: I1203 00:34:10.305961 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-5qxfs" Dec 03 00:34:10 crc kubenswrapper[4953]: I1203 00:34:10.319329 4953 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-5qxfs"] Dec 03 00:34:10 crc kubenswrapper[4953]: I1203 00:34:10.488685 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-846sz\" (UniqueName: \"kubernetes.io/projected/aae059a0-2ee5-4dee-9b0c-154a358dceda-kube-api-access-846sz\") pod \"community-operators-5qxfs\" (UID: \"aae059a0-2ee5-4dee-9b0c-154a358dceda\") " pod="openshift-marketplace/community-operators-5qxfs" Dec 03 00:34:10 crc kubenswrapper[4953]: I1203 00:34:10.488779 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aae059a0-2ee5-4dee-9b0c-154a358dceda-utilities\") pod \"community-operators-5qxfs\" (UID: \"aae059a0-2ee5-4dee-9b0c-154a358dceda\") " pod="openshift-marketplace/community-operators-5qxfs" Dec 03 00:34:10 crc kubenswrapper[4953]: I1203 00:34:10.488827 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aae059a0-2ee5-4dee-9b0c-154a358dceda-catalog-content\") pod \"community-operators-5qxfs\" (UID: \"aae059a0-2ee5-4dee-9b0c-154a358dceda\") " pod="openshift-marketplace/community-operators-5qxfs" Dec 03 00:34:10 crc kubenswrapper[4953]: I1203 00:34:10.505003 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/service-telemetry-operator-bundle-2-build" event={"ID":"3fd62fed-256b-4d94-8757-9271988b9119","Type":"ContainerStarted","Data":"f7557fd50527bb568d76950664cc829afde51169899e03a7469190d9bc7d23d2"} Dec 03 00:34:10 crc kubenswrapper[4953]: I1203 00:34:10.538096 4953 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="service-telemetry/service-telemetry-operator-bundle-2-build" podStartSLOduration=5.538068236 podStartE2EDuration="5.538068236s" podCreationTimestamp="2025-12-03 00:34:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 00:34:10.533711799 +0000 UTC m=+1639.517992592" watchObservedRunningTime="2025-12-03 00:34:10.538068236 +0000 UTC m=+1639.522349029" Dec 03 00:34:10 crc kubenswrapper[4953]: I1203 00:34:10.590311 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-846sz\" (UniqueName: \"kubernetes.io/projected/aae059a0-2ee5-4dee-9b0c-154a358dceda-kube-api-access-846sz\") pod \"community-operators-5qxfs\" (UID: \"aae059a0-2ee5-4dee-9b0c-154a358dceda\") " pod="openshift-marketplace/community-operators-5qxfs" Dec 03 00:34:10 crc kubenswrapper[4953]: I1203 00:34:10.590411 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aae059a0-2ee5-4dee-9b0c-154a358dceda-utilities\") pod \"community-operators-5qxfs\" (UID: \"aae059a0-2ee5-4dee-9b0c-154a358dceda\") " pod="openshift-marketplace/community-operators-5qxfs" Dec 03 00:34:10 crc kubenswrapper[4953]: I1203 00:34:10.590451 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aae059a0-2ee5-4dee-9b0c-154a358dceda-catalog-content\") pod \"community-operators-5qxfs\" (UID: \"aae059a0-2ee5-4dee-9b0c-154a358dceda\") " pod="openshift-marketplace/community-operators-5qxfs" Dec 03 00:34:10 crc kubenswrapper[4953]: I1203 00:34:10.591081 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aae059a0-2ee5-4dee-9b0c-154a358dceda-catalog-content\") pod \"community-operators-5qxfs\" (UID: \"aae059a0-2ee5-4dee-9b0c-154a358dceda\") " pod="openshift-marketplace/community-operators-5qxfs" Dec 03 00:34:10 crc kubenswrapper[4953]: I1203 00:34:10.591788 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aae059a0-2ee5-4dee-9b0c-154a358dceda-utilities\") pod \"community-operators-5qxfs\" (UID: \"aae059a0-2ee5-4dee-9b0c-154a358dceda\") " pod="openshift-marketplace/community-operators-5qxfs" Dec 03 00:34:10 crc kubenswrapper[4953]: I1203 00:34:10.612762 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-846sz\" (UniqueName: \"kubernetes.io/projected/aae059a0-2ee5-4dee-9b0c-154a358dceda-kube-api-access-846sz\") pod \"community-operators-5qxfs\" (UID: \"aae059a0-2ee5-4dee-9b0c-154a358dceda\") " pod="openshift-marketplace/community-operators-5qxfs" Dec 03 00:34:10 crc kubenswrapper[4953]: I1203 00:34:10.669363 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-5qxfs" Dec 03 00:34:10 crc kubenswrapper[4953]: I1203 00:34:10.958506 4953 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-5qxfs"] Dec 03 00:34:11 crc kubenswrapper[4953]: I1203 00:34:11.515619 4953 generic.go:334] "Generic (PLEG): container finished" podID="aae059a0-2ee5-4dee-9b0c-154a358dceda" containerID="e526fbd002ca75bcecc7b726353988108b4da843e4d7be29027defec275563dc" exitCode=0 Dec 03 00:34:11 crc kubenswrapper[4953]: I1203 00:34:11.515700 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5qxfs" event={"ID":"aae059a0-2ee5-4dee-9b0c-154a358dceda","Type":"ContainerDied","Data":"e526fbd002ca75bcecc7b726353988108b4da843e4d7be29027defec275563dc"} Dec 03 00:34:11 crc kubenswrapper[4953]: I1203 00:34:11.515770 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5qxfs" event={"ID":"aae059a0-2ee5-4dee-9b0c-154a358dceda","Type":"ContainerStarted","Data":"b8fec301fe33a1f62600b1d85d9741e2e3c27fcb1d5490ee8057ba410ceb3066"} Dec 03 00:34:13 crc kubenswrapper[4953]: I1203 00:34:13.533578 4953 generic.go:334] "Generic (PLEG): container finished" podID="3fd62fed-256b-4d94-8757-9271988b9119" containerID="f7557fd50527bb568d76950664cc829afde51169899e03a7469190d9bc7d23d2" exitCode=0 Dec 03 00:34:13 crc kubenswrapper[4953]: I1203 00:34:13.534229 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/service-telemetry-operator-bundle-2-build" event={"ID":"3fd62fed-256b-4d94-8757-9271988b9119","Type":"ContainerDied","Data":"f7557fd50527bb568d76950664cc829afde51169899e03a7469190d9bc7d23d2"} Dec 03 00:34:13 crc kubenswrapper[4953]: I1203 00:34:13.537599 4953 generic.go:334] "Generic (PLEG): container finished" podID="aae059a0-2ee5-4dee-9b0c-154a358dceda" containerID="d979356ce713cf58cf310e70f827e19c3cb0ef0169ddc2d50351feb628b58067" exitCode=0 Dec 03 00:34:13 crc kubenswrapper[4953]: I1203 00:34:13.537667 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5qxfs" event={"ID":"aae059a0-2ee5-4dee-9b0c-154a358dceda","Type":"ContainerDied","Data":"d979356ce713cf58cf310e70f827e19c3cb0ef0169ddc2d50351feb628b58067"} Dec 03 00:34:14 crc kubenswrapper[4953]: I1203 00:34:14.547954 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5qxfs" event={"ID":"aae059a0-2ee5-4dee-9b0c-154a358dceda","Type":"ContainerStarted","Data":"61039323d67c2756708f022e475445f997dd5cbca40aada01fdd770bff8a3cf0"} Dec 03 00:34:14 crc kubenswrapper[4953]: I1203 00:34:14.572024 4953 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-5qxfs" podStartSLOduration=2.129477113 podStartE2EDuration="4.57199918s" podCreationTimestamp="2025-12-03 00:34:10 +0000 UTC" firstStartedPulling="2025-12-03 00:34:11.518747737 +0000 UTC m=+1640.503028520" lastFinishedPulling="2025-12-03 00:34:13.961269804 +0000 UTC m=+1642.945550587" observedRunningTime="2025-12-03 00:34:14.565967672 +0000 UTC m=+1643.550248455" watchObservedRunningTime="2025-12-03 00:34:14.57199918 +0000 UTC m=+1643.556279963" Dec 03 00:34:14 crc kubenswrapper[4953]: I1203 00:34:14.821368 4953 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/service-telemetry-operator-bundle-2-build" Dec 03 00:34:14 crc kubenswrapper[4953]: I1203 00:34:14.962994 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mbj7m\" (UniqueName: \"kubernetes.io/projected/3fd62fed-256b-4d94-8757-9271988b9119-kube-api-access-mbj7m\") pod \"3fd62fed-256b-4d94-8757-9271988b9119\" (UID: \"3fd62fed-256b-4d94-8757-9271988b9119\") " Dec 03 00:34:14 crc kubenswrapper[4953]: I1203 00:34:14.963089 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/3fd62fed-256b-4d94-8757-9271988b9119-build-system-configs\") pod \"3fd62fed-256b-4d94-8757-9271988b9119\" (UID: \"3fd62fed-256b-4d94-8757-9271988b9119\") " Dec 03 00:34:14 crc kubenswrapper[4953]: I1203 00:34:14.963130 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/3fd62fed-256b-4d94-8757-9271988b9119-container-storage-root\") pod \"3fd62fed-256b-4d94-8757-9271988b9119\" (UID: \"3fd62fed-256b-4d94-8757-9271988b9119\") " Dec 03 00:34:14 crc kubenswrapper[4953]: I1203 00:34:14.963171 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/3fd62fed-256b-4d94-8757-9271988b9119-node-pullsecrets\") pod \"3fd62fed-256b-4d94-8757-9271988b9119\" (UID: \"3fd62fed-256b-4d94-8757-9271988b9119\") " Dec 03 00:34:14 crc kubenswrapper[4953]: I1203 00:34:14.963199 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/3fd62fed-256b-4d94-8757-9271988b9119-buildcachedir\") pod \"3fd62fed-256b-4d94-8757-9271988b9119\" (UID: \"3fd62fed-256b-4d94-8757-9271988b9119\") " Dec 03 00:34:14 crc kubenswrapper[4953]: I1203 00:34:14.963219 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/3fd62fed-256b-4d94-8757-9271988b9119-build-proxy-ca-bundles\") pod \"3fd62fed-256b-4d94-8757-9271988b9119\" (UID: \"3fd62fed-256b-4d94-8757-9271988b9119\") " Dec 03 00:34:14 crc kubenswrapper[4953]: I1203 00:34:14.963297 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/3fd62fed-256b-4d94-8757-9271988b9119-build-ca-bundles\") pod \"3fd62fed-256b-4d94-8757-9271988b9119\" (UID: \"3fd62fed-256b-4d94-8757-9271988b9119\") " Dec 03 00:34:14 crc kubenswrapper[4953]: I1203 00:34:14.964070 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"builder-dockercfg-9prwd-push\" (UniqueName: \"kubernetes.io/secret/3fd62fed-256b-4d94-8757-9271988b9119-builder-dockercfg-9prwd-push\") pod \"3fd62fed-256b-4d94-8757-9271988b9119\" (UID: \"3fd62fed-256b-4d94-8757-9271988b9119\") " Dec 03 00:34:14 crc kubenswrapper[4953]: I1203 00:34:14.964107 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"builder-dockercfg-9prwd-pull\" (UniqueName: \"kubernetes.io/secret/3fd62fed-256b-4d94-8757-9271988b9119-builder-dockercfg-9prwd-pull\") pod \"3fd62fed-256b-4d94-8757-9271988b9119\" (UID: \"3fd62fed-256b-4d94-8757-9271988b9119\") " Dec 03 00:34:14 crc kubenswrapper[4953]: I1203 00:34:14.964132 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/3fd62fed-256b-4d94-8757-9271988b9119-buildworkdir\") pod \"3fd62fed-256b-4d94-8757-9271988b9119\" (UID: \"3fd62fed-256b-4d94-8757-9271988b9119\") " Dec 03 00:34:14 crc kubenswrapper[4953]: I1203 00:34:14.964169 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/3fd62fed-256b-4d94-8757-9271988b9119-container-storage-run\") pod \"3fd62fed-256b-4d94-8757-9271988b9119\" (UID: \"3fd62fed-256b-4d94-8757-9271988b9119\") " Dec 03 00:34:14 crc kubenswrapper[4953]: I1203 00:34:14.964226 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/3fd62fed-256b-4d94-8757-9271988b9119-build-blob-cache\") pod \"3fd62fed-256b-4d94-8757-9271988b9119\" (UID: \"3fd62fed-256b-4d94-8757-9271988b9119\") " Dec 03 00:34:14 crc kubenswrapper[4953]: I1203 00:34:14.963612 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3fd62fed-256b-4d94-8757-9271988b9119-buildcachedir" (OuterVolumeSpecName: "buildcachedir") pod "3fd62fed-256b-4d94-8757-9271988b9119" (UID: "3fd62fed-256b-4d94-8757-9271988b9119"). InnerVolumeSpecName "buildcachedir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 00:34:14 crc kubenswrapper[4953]: I1203 00:34:14.963679 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3fd62fed-256b-4d94-8757-9271988b9119-node-pullsecrets" (OuterVolumeSpecName: "node-pullsecrets") pod "3fd62fed-256b-4d94-8757-9271988b9119" (UID: "3fd62fed-256b-4d94-8757-9271988b9119"). InnerVolumeSpecName "node-pullsecrets". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 00:34:14 crc kubenswrapper[4953]: I1203 00:34:14.964687 4953 reconciler_common.go:293] "Volume detached for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/3fd62fed-256b-4d94-8757-9271988b9119-node-pullsecrets\") on node \"crc\" DevicePath \"\"" Dec 03 00:34:14 crc kubenswrapper[4953]: I1203 00:34:14.964708 4953 reconciler_common.go:293] "Volume detached for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/3fd62fed-256b-4d94-8757-9271988b9119-buildcachedir\") on node \"crc\" DevicePath \"\"" Dec 03 00:34:14 crc kubenswrapper[4953]: I1203 00:34:14.964754 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3fd62fed-256b-4d94-8757-9271988b9119-build-proxy-ca-bundles" (OuterVolumeSpecName: "build-proxy-ca-bundles") pod "3fd62fed-256b-4d94-8757-9271988b9119" (UID: "3fd62fed-256b-4d94-8757-9271988b9119"). InnerVolumeSpecName "build-proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:34:14 crc kubenswrapper[4953]: I1203 00:34:14.965361 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3fd62fed-256b-4d94-8757-9271988b9119-build-system-configs" (OuterVolumeSpecName: "build-system-configs") pod "3fd62fed-256b-4d94-8757-9271988b9119" (UID: "3fd62fed-256b-4d94-8757-9271988b9119"). InnerVolumeSpecName "build-system-configs". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:34:14 crc kubenswrapper[4953]: I1203 00:34:14.966640 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3fd62fed-256b-4d94-8757-9271988b9119-container-storage-run" (OuterVolumeSpecName: "container-storage-run") pod "3fd62fed-256b-4d94-8757-9271988b9119" (UID: "3fd62fed-256b-4d94-8757-9271988b9119"). InnerVolumeSpecName "container-storage-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 00:34:14 crc kubenswrapper[4953]: I1203 00:34:14.968678 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3fd62fed-256b-4d94-8757-9271988b9119-buildworkdir" (OuterVolumeSpecName: "buildworkdir") pod "3fd62fed-256b-4d94-8757-9271988b9119" (UID: "3fd62fed-256b-4d94-8757-9271988b9119"). InnerVolumeSpecName "buildworkdir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 00:34:14 crc kubenswrapper[4953]: I1203 00:34:14.970340 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3fd62fed-256b-4d94-8757-9271988b9119-build-blob-cache" (OuterVolumeSpecName: "build-blob-cache") pod "3fd62fed-256b-4d94-8757-9271988b9119" (UID: "3fd62fed-256b-4d94-8757-9271988b9119"). InnerVolumeSpecName "build-blob-cache". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 00:34:14 crc kubenswrapper[4953]: I1203 00:34:14.970368 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3fd62fed-256b-4d94-8757-9271988b9119-builder-dockercfg-9prwd-pull" (OuterVolumeSpecName: "builder-dockercfg-9prwd-pull") pod "3fd62fed-256b-4d94-8757-9271988b9119" (UID: "3fd62fed-256b-4d94-8757-9271988b9119"). InnerVolumeSpecName "builder-dockercfg-9prwd-pull". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:34:14 crc kubenswrapper[4953]: I1203 00:34:14.970590 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3fd62fed-256b-4d94-8757-9271988b9119-builder-dockercfg-9prwd-push" (OuterVolumeSpecName: "builder-dockercfg-9prwd-push") pod "3fd62fed-256b-4d94-8757-9271988b9119" (UID: "3fd62fed-256b-4d94-8757-9271988b9119"). InnerVolumeSpecName "builder-dockercfg-9prwd-push". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:34:14 crc kubenswrapper[4953]: I1203 00:34:14.970664 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3fd62fed-256b-4d94-8757-9271988b9119-kube-api-access-mbj7m" (OuterVolumeSpecName: "kube-api-access-mbj7m") pod "3fd62fed-256b-4d94-8757-9271988b9119" (UID: "3fd62fed-256b-4d94-8757-9271988b9119"). InnerVolumeSpecName "kube-api-access-mbj7m". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:34:14 crc kubenswrapper[4953]: I1203 00:34:14.973950 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3fd62fed-256b-4d94-8757-9271988b9119-container-storage-root" (OuterVolumeSpecName: "container-storage-root") pod "3fd62fed-256b-4d94-8757-9271988b9119" (UID: "3fd62fed-256b-4d94-8757-9271988b9119"). InnerVolumeSpecName "container-storage-root". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 00:34:14 crc kubenswrapper[4953]: I1203 00:34:14.982183 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3fd62fed-256b-4d94-8757-9271988b9119-build-ca-bundles" (OuterVolumeSpecName: "build-ca-bundles") pod "3fd62fed-256b-4d94-8757-9271988b9119" (UID: "3fd62fed-256b-4d94-8757-9271988b9119"). InnerVolumeSpecName "build-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:34:15 crc kubenswrapper[4953]: I1203 00:34:15.067399 4953 reconciler_common.go:293] "Volume detached for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/3fd62fed-256b-4d94-8757-9271988b9119-build-ca-bundles\") on node \"crc\" DevicePath \"\"" Dec 03 00:34:15 crc kubenswrapper[4953]: I1203 00:34:15.067481 4953 reconciler_common.go:293] "Volume detached for volume \"builder-dockercfg-9prwd-push\" (UniqueName: \"kubernetes.io/secret/3fd62fed-256b-4d94-8757-9271988b9119-builder-dockercfg-9prwd-push\") on node \"crc\" DevicePath \"\"" Dec 03 00:34:15 crc kubenswrapper[4953]: I1203 00:34:15.067503 4953 reconciler_common.go:293] "Volume detached for volume \"builder-dockercfg-9prwd-pull\" (UniqueName: \"kubernetes.io/secret/3fd62fed-256b-4d94-8757-9271988b9119-builder-dockercfg-9prwd-pull\") on node \"crc\" DevicePath \"\"" Dec 03 00:34:15 crc kubenswrapper[4953]: I1203 00:34:15.067520 4953 reconciler_common.go:293] "Volume detached for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/3fd62fed-256b-4d94-8757-9271988b9119-buildworkdir\") on node \"crc\" DevicePath \"\"" Dec 03 00:34:15 crc kubenswrapper[4953]: I1203 00:34:15.067533 4953 reconciler_common.go:293] "Volume detached for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/3fd62fed-256b-4d94-8757-9271988b9119-container-storage-run\") on node \"crc\" DevicePath \"\"" Dec 03 00:34:15 crc kubenswrapper[4953]: I1203 00:34:15.067549 4953 reconciler_common.go:293] "Volume detached for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/3fd62fed-256b-4d94-8757-9271988b9119-build-blob-cache\") on node \"crc\" DevicePath \"\"" Dec 03 00:34:15 crc kubenswrapper[4953]: I1203 00:34:15.067564 4953 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mbj7m\" (UniqueName: \"kubernetes.io/projected/3fd62fed-256b-4d94-8757-9271988b9119-kube-api-access-mbj7m\") on node \"crc\" DevicePath \"\"" Dec 03 00:34:15 crc kubenswrapper[4953]: I1203 00:34:15.067577 4953 reconciler_common.go:293] "Volume detached for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/3fd62fed-256b-4d94-8757-9271988b9119-build-system-configs\") on node \"crc\" DevicePath \"\"" Dec 03 00:34:15 crc kubenswrapper[4953]: I1203 00:34:15.067589 4953 reconciler_common.go:293] "Volume detached for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/3fd62fed-256b-4d94-8757-9271988b9119-container-storage-root\") on node \"crc\" DevicePath \"\"" Dec 03 00:34:15 crc kubenswrapper[4953]: I1203 00:34:15.067601 4953 reconciler_common.go:293] "Volume detached for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/3fd62fed-256b-4d94-8757-9271988b9119-build-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Dec 03 00:34:15 crc kubenswrapper[4953]: I1203 00:34:15.557564 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/service-telemetry-operator-bundle-2-build" event={"ID":"3fd62fed-256b-4d94-8757-9271988b9119","Type":"ContainerDied","Data":"56aa65cc8e5b3f8376d5dd43fa360a6fb8221d2ca626286ecb48e1009c913592"} Dec 03 00:34:15 crc kubenswrapper[4953]: I1203 00:34:15.557635 4953 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/service-telemetry-operator-bundle-2-build" Dec 03 00:34:15 crc kubenswrapper[4953]: I1203 00:34:15.557653 4953 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="56aa65cc8e5b3f8376d5dd43fa360a6fb8221d2ca626286ecb48e1009c913592" Dec 03 00:34:18 crc kubenswrapper[4953]: I1203 00:34:18.655423 4953 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["service-telemetry/smart-gateway-operator-bundle-1-build"] Dec 03 00:34:18 crc kubenswrapper[4953]: E1203 00:34:18.656439 4953 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3fd62fed-256b-4d94-8757-9271988b9119" containerName="manage-dockerfile" Dec 03 00:34:18 crc kubenswrapper[4953]: I1203 00:34:18.656461 4953 state_mem.go:107] "Deleted CPUSet assignment" podUID="3fd62fed-256b-4d94-8757-9271988b9119" containerName="manage-dockerfile" Dec 03 00:34:18 crc kubenswrapper[4953]: E1203 00:34:18.656484 4953 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3fd62fed-256b-4d94-8757-9271988b9119" containerName="docker-build" Dec 03 00:34:18 crc kubenswrapper[4953]: I1203 00:34:18.656494 4953 state_mem.go:107] "Deleted CPUSet assignment" podUID="3fd62fed-256b-4d94-8757-9271988b9119" containerName="docker-build" Dec 03 00:34:18 crc kubenswrapper[4953]: E1203 00:34:18.656506 4953 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3fd62fed-256b-4d94-8757-9271988b9119" containerName="git-clone" Dec 03 00:34:18 crc kubenswrapper[4953]: I1203 00:34:18.656517 4953 state_mem.go:107] "Deleted CPUSet assignment" podUID="3fd62fed-256b-4d94-8757-9271988b9119" containerName="git-clone" Dec 03 00:34:18 crc kubenswrapper[4953]: I1203 00:34:18.656720 4953 memory_manager.go:354] "RemoveStaleState removing state" podUID="3fd62fed-256b-4d94-8757-9271988b9119" containerName="docker-build" Dec 03 00:34:18 crc kubenswrapper[4953]: I1203 00:34:18.658032 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/smart-gateway-operator-bundle-1-build" Dec 03 00:34:18 crc kubenswrapper[4953]: I1203 00:34:18.667514 4953 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"builder-dockercfg-9prwd" Dec 03 00:34:18 crc kubenswrapper[4953]: I1203 00:34:18.670096 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"smart-gateway-operator-bundle-1-sys-config" Dec 03 00:34:18 crc kubenswrapper[4953]: I1203 00:34:18.670220 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"smart-gateway-operator-bundle-1-ca" Dec 03 00:34:18 crc kubenswrapper[4953]: I1203 00:34:18.670360 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"smart-gateway-operator-bundle-1-global-ca" Dec 03 00:34:18 crc kubenswrapper[4953]: I1203 00:34:18.670886 4953 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/smart-gateway-operator-bundle-1-build"] Dec 03 00:34:18 crc kubenswrapper[4953]: I1203 00:34:18.827045 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/c668590f-b40e-4c3c-9fa5-33cfcb689fc2-node-pullsecrets\") pod \"smart-gateway-operator-bundle-1-build\" (UID: \"c668590f-b40e-4c3c-9fa5-33cfcb689fc2\") " pod="service-telemetry/smart-gateway-operator-bundle-1-build" Dec 03 00:34:18 crc kubenswrapper[4953]: I1203 00:34:18.827125 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"builder-dockercfg-9prwd-push\" (UniqueName: \"kubernetes.io/secret/c668590f-b40e-4c3c-9fa5-33cfcb689fc2-builder-dockercfg-9prwd-push\") pod \"smart-gateway-operator-bundle-1-build\" (UID: \"c668590f-b40e-4c3c-9fa5-33cfcb689fc2\") " pod="service-telemetry/smart-gateway-operator-bundle-1-build" Dec 03 00:34:18 crc kubenswrapper[4953]: I1203 00:34:18.827306 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/c668590f-b40e-4c3c-9fa5-33cfcb689fc2-build-proxy-ca-bundles\") pod \"smart-gateway-operator-bundle-1-build\" (UID: \"c668590f-b40e-4c3c-9fa5-33cfcb689fc2\") " pod="service-telemetry/smart-gateway-operator-bundle-1-build" Dec 03 00:34:18 crc kubenswrapper[4953]: I1203 00:34:18.827554 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/c668590f-b40e-4c3c-9fa5-33cfcb689fc2-build-system-configs\") pod \"smart-gateway-operator-bundle-1-build\" (UID: \"c668590f-b40e-4c3c-9fa5-33cfcb689fc2\") " pod="service-telemetry/smart-gateway-operator-bundle-1-build" Dec 03 00:34:18 crc kubenswrapper[4953]: I1203 00:34:18.827634 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/c668590f-b40e-4c3c-9fa5-33cfcb689fc2-container-storage-run\") pod \"smart-gateway-operator-bundle-1-build\" (UID: \"c668590f-b40e-4c3c-9fa5-33cfcb689fc2\") " pod="service-telemetry/smart-gateway-operator-bundle-1-build" Dec 03 00:34:18 crc kubenswrapper[4953]: I1203 00:34:18.827663 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/c668590f-b40e-4c3c-9fa5-33cfcb689fc2-build-blob-cache\") pod \"smart-gateway-operator-bundle-1-build\" (UID: \"c668590f-b40e-4c3c-9fa5-33cfcb689fc2\") " pod="service-telemetry/smart-gateway-operator-bundle-1-build" Dec 03 00:34:18 crc kubenswrapper[4953]: I1203 00:34:18.827734 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"builder-dockercfg-9prwd-pull\" (UniqueName: \"kubernetes.io/secret/c668590f-b40e-4c3c-9fa5-33cfcb689fc2-builder-dockercfg-9prwd-pull\") pod \"smart-gateway-operator-bundle-1-build\" (UID: \"c668590f-b40e-4c3c-9fa5-33cfcb689fc2\") " pod="service-telemetry/smart-gateway-operator-bundle-1-build" Dec 03 00:34:18 crc kubenswrapper[4953]: I1203 00:34:18.827776 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/c668590f-b40e-4c3c-9fa5-33cfcb689fc2-buildworkdir\") pod \"smart-gateway-operator-bundle-1-build\" (UID: \"c668590f-b40e-4c3c-9fa5-33cfcb689fc2\") " pod="service-telemetry/smart-gateway-operator-bundle-1-build" Dec 03 00:34:18 crc kubenswrapper[4953]: I1203 00:34:18.827802 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/c668590f-b40e-4c3c-9fa5-33cfcb689fc2-container-storage-root\") pod \"smart-gateway-operator-bundle-1-build\" (UID: \"c668590f-b40e-4c3c-9fa5-33cfcb689fc2\") " pod="service-telemetry/smart-gateway-operator-bundle-1-build" Dec 03 00:34:18 crc kubenswrapper[4953]: I1203 00:34:18.827829 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/c668590f-b40e-4c3c-9fa5-33cfcb689fc2-build-ca-bundles\") pod \"smart-gateway-operator-bundle-1-build\" (UID: \"c668590f-b40e-4c3c-9fa5-33cfcb689fc2\") " pod="service-telemetry/smart-gateway-operator-bundle-1-build" Dec 03 00:34:18 crc kubenswrapper[4953]: I1203 00:34:18.827921 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bwwlt\" (UniqueName: \"kubernetes.io/projected/c668590f-b40e-4c3c-9fa5-33cfcb689fc2-kube-api-access-bwwlt\") pod \"smart-gateway-operator-bundle-1-build\" (UID: \"c668590f-b40e-4c3c-9fa5-33cfcb689fc2\") " pod="service-telemetry/smart-gateway-operator-bundle-1-build" Dec 03 00:34:18 crc kubenswrapper[4953]: I1203 00:34:18.827963 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/c668590f-b40e-4c3c-9fa5-33cfcb689fc2-buildcachedir\") pod \"smart-gateway-operator-bundle-1-build\" (UID: \"c668590f-b40e-4c3c-9fa5-33cfcb689fc2\") " pod="service-telemetry/smart-gateway-operator-bundle-1-build" Dec 03 00:34:18 crc kubenswrapper[4953]: I1203 00:34:18.929492 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bwwlt\" (UniqueName: \"kubernetes.io/projected/c668590f-b40e-4c3c-9fa5-33cfcb689fc2-kube-api-access-bwwlt\") pod \"smart-gateway-operator-bundle-1-build\" (UID: \"c668590f-b40e-4c3c-9fa5-33cfcb689fc2\") " pod="service-telemetry/smart-gateway-operator-bundle-1-build" Dec 03 00:34:18 crc kubenswrapper[4953]: I1203 00:34:18.929576 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/c668590f-b40e-4c3c-9fa5-33cfcb689fc2-buildcachedir\") pod \"smart-gateway-operator-bundle-1-build\" (UID: \"c668590f-b40e-4c3c-9fa5-33cfcb689fc2\") " pod="service-telemetry/smart-gateway-operator-bundle-1-build" Dec 03 00:34:18 crc kubenswrapper[4953]: I1203 00:34:18.929635 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/c668590f-b40e-4c3c-9fa5-33cfcb689fc2-node-pullsecrets\") pod \"smart-gateway-operator-bundle-1-build\" (UID: \"c668590f-b40e-4c3c-9fa5-33cfcb689fc2\") " pod="service-telemetry/smart-gateway-operator-bundle-1-build" Dec 03 00:34:18 crc kubenswrapper[4953]: I1203 00:34:18.929691 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"builder-dockercfg-9prwd-push\" (UniqueName: \"kubernetes.io/secret/c668590f-b40e-4c3c-9fa5-33cfcb689fc2-builder-dockercfg-9prwd-push\") pod \"smart-gateway-operator-bundle-1-build\" (UID: \"c668590f-b40e-4c3c-9fa5-33cfcb689fc2\") " pod="service-telemetry/smart-gateway-operator-bundle-1-build" Dec 03 00:34:18 crc kubenswrapper[4953]: I1203 00:34:18.929759 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/c668590f-b40e-4c3c-9fa5-33cfcb689fc2-build-proxy-ca-bundles\") pod \"smart-gateway-operator-bundle-1-build\" (UID: \"c668590f-b40e-4c3c-9fa5-33cfcb689fc2\") " pod="service-telemetry/smart-gateway-operator-bundle-1-build" Dec 03 00:34:18 crc kubenswrapper[4953]: I1203 00:34:18.929852 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/c668590f-b40e-4c3c-9fa5-33cfcb689fc2-node-pullsecrets\") pod \"smart-gateway-operator-bundle-1-build\" (UID: \"c668590f-b40e-4c3c-9fa5-33cfcb689fc2\") " pod="service-telemetry/smart-gateway-operator-bundle-1-build" Dec 03 00:34:18 crc kubenswrapper[4953]: I1203 00:34:18.929882 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/c668590f-b40e-4c3c-9fa5-33cfcb689fc2-build-system-configs\") pod \"smart-gateway-operator-bundle-1-build\" (UID: \"c668590f-b40e-4c3c-9fa5-33cfcb689fc2\") " pod="service-telemetry/smart-gateway-operator-bundle-1-build" Dec 03 00:34:18 crc kubenswrapper[4953]: I1203 00:34:18.929995 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/c668590f-b40e-4c3c-9fa5-33cfcb689fc2-container-storage-run\") pod \"smart-gateway-operator-bundle-1-build\" (UID: \"c668590f-b40e-4c3c-9fa5-33cfcb689fc2\") " pod="service-telemetry/smart-gateway-operator-bundle-1-build" Dec 03 00:34:18 crc kubenswrapper[4953]: I1203 00:34:18.930029 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/c668590f-b40e-4c3c-9fa5-33cfcb689fc2-build-blob-cache\") pod \"smart-gateway-operator-bundle-1-build\" (UID: \"c668590f-b40e-4c3c-9fa5-33cfcb689fc2\") " pod="service-telemetry/smart-gateway-operator-bundle-1-build" Dec 03 00:34:18 crc kubenswrapper[4953]: I1203 00:34:18.930083 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"builder-dockercfg-9prwd-pull\" (UniqueName: \"kubernetes.io/secret/c668590f-b40e-4c3c-9fa5-33cfcb689fc2-builder-dockercfg-9prwd-pull\") pod \"smart-gateway-operator-bundle-1-build\" (UID: \"c668590f-b40e-4c3c-9fa5-33cfcb689fc2\") " pod="service-telemetry/smart-gateway-operator-bundle-1-build" Dec 03 00:34:18 crc kubenswrapper[4953]: I1203 00:34:18.930123 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/c668590f-b40e-4c3c-9fa5-33cfcb689fc2-buildworkdir\") pod \"smart-gateway-operator-bundle-1-build\" (UID: \"c668590f-b40e-4c3c-9fa5-33cfcb689fc2\") " pod="service-telemetry/smart-gateway-operator-bundle-1-build" Dec 03 00:34:18 crc kubenswrapper[4953]: I1203 00:34:18.930143 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/c668590f-b40e-4c3c-9fa5-33cfcb689fc2-container-storage-root\") pod \"smart-gateway-operator-bundle-1-build\" (UID: \"c668590f-b40e-4c3c-9fa5-33cfcb689fc2\") " pod="service-telemetry/smart-gateway-operator-bundle-1-build" Dec 03 00:34:18 crc kubenswrapper[4953]: I1203 00:34:18.930189 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/c668590f-b40e-4c3c-9fa5-33cfcb689fc2-build-ca-bundles\") pod \"smart-gateway-operator-bundle-1-build\" (UID: \"c668590f-b40e-4c3c-9fa5-33cfcb689fc2\") " pod="service-telemetry/smart-gateway-operator-bundle-1-build" Dec 03 00:34:18 crc kubenswrapper[4953]: I1203 00:34:18.929760 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/c668590f-b40e-4c3c-9fa5-33cfcb689fc2-buildcachedir\") pod \"smart-gateway-operator-bundle-1-build\" (UID: \"c668590f-b40e-4c3c-9fa5-33cfcb689fc2\") " pod="service-telemetry/smart-gateway-operator-bundle-1-build" Dec 03 00:34:18 crc kubenswrapper[4953]: I1203 00:34:18.931104 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/c668590f-b40e-4c3c-9fa5-33cfcb689fc2-build-system-configs\") pod \"smart-gateway-operator-bundle-1-build\" (UID: \"c668590f-b40e-4c3c-9fa5-33cfcb689fc2\") " pod="service-telemetry/smart-gateway-operator-bundle-1-build" Dec 03 00:34:18 crc kubenswrapper[4953]: I1203 00:34:18.931212 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/c668590f-b40e-4c3c-9fa5-33cfcb689fc2-build-blob-cache\") pod \"smart-gateway-operator-bundle-1-build\" (UID: \"c668590f-b40e-4c3c-9fa5-33cfcb689fc2\") " pod="service-telemetry/smart-gateway-operator-bundle-1-build" Dec 03 00:34:18 crc kubenswrapper[4953]: I1203 00:34:18.931235 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/c668590f-b40e-4c3c-9fa5-33cfcb689fc2-container-storage-root\") pod \"smart-gateway-operator-bundle-1-build\" (UID: \"c668590f-b40e-4c3c-9fa5-33cfcb689fc2\") " pod="service-telemetry/smart-gateway-operator-bundle-1-build" Dec 03 00:34:18 crc kubenswrapper[4953]: I1203 00:34:18.931354 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/c668590f-b40e-4c3c-9fa5-33cfcb689fc2-build-ca-bundles\") pod \"smart-gateway-operator-bundle-1-build\" (UID: \"c668590f-b40e-4c3c-9fa5-33cfcb689fc2\") " pod="service-telemetry/smart-gateway-operator-bundle-1-build" Dec 03 00:34:18 crc kubenswrapper[4953]: I1203 00:34:18.931906 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/c668590f-b40e-4c3c-9fa5-33cfcb689fc2-build-proxy-ca-bundles\") pod \"smart-gateway-operator-bundle-1-build\" (UID: \"c668590f-b40e-4c3c-9fa5-33cfcb689fc2\") " pod="service-telemetry/smart-gateway-operator-bundle-1-build" Dec 03 00:34:18 crc kubenswrapper[4953]: I1203 00:34:18.932511 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/c668590f-b40e-4c3c-9fa5-33cfcb689fc2-container-storage-run\") pod \"smart-gateway-operator-bundle-1-build\" (UID: \"c668590f-b40e-4c3c-9fa5-33cfcb689fc2\") " pod="service-telemetry/smart-gateway-operator-bundle-1-build" Dec 03 00:34:18 crc kubenswrapper[4953]: I1203 00:34:18.932638 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/c668590f-b40e-4c3c-9fa5-33cfcb689fc2-buildworkdir\") pod \"smart-gateway-operator-bundle-1-build\" (UID: \"c668590f-b40e-4c3c-9fa5-33cfcb689fc2\") " pod="service-telemetry/smart-gateway-operator-bundle-1-build" Dec 03 00:34:18 crc kubenswrapper[4953]: I1203 00:34:18.939361 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"builder-dockercfg-9prwd-pull\" (UniqueName: \"kubernetes.io/secret/c668590f-b40e-4c3c-9fa5-33cfcb689fc2-builder-dockercfg-9prwd-pull\") pod \"smart-gateway-operator-bundle-1-build\" (UID: \"c668590f-b40e-4c3c-9fa5-33cfcb689fc2\") " pod="service-telemetry/smart-gateway-operator-bundle-1-build" Dec 03 00:34:18 crc kubenswrapper[4953]: I1203 00:34:18.940260 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"builder-dockercfg-9prwd-push\" (UniqueName: \"kubernetes.io/secret/c668590f-b40e-4c3c-9fa5-33cfcb689fc2-builder-dockercfg-9prwd-push\") pod \"smart-gateway-operator-bundle-1-build\" (UID: \"c668590f-b40e-4c3c-9fa5-33cfcb689fc2\") " pod="service-telemetry/smart-gateway-operator-bundle-1-build" Dec 03 00:34:18 crc kubenswrapper[4953]: I1203 00:34:18.944930 4953 patch_prober.go:28] interesting pod/machine-config-daemon-6dfbn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 00:34:18 crc kubenswrapper[4953]: I1203 00:34:18.945042 4953 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-6dfbn" podUID="53ffb1e9-9a2e-4869-80d7-776f4aa7ef8b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 00:34:18 crc kubenswrapper[4953]: I1203 00:34:18.945119 4953 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-6dfbn" Dec 03 00:34:18 crc kubenswrapper[4953]: I1203 00:34:18.946480 4953 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"150ffc325f215c6da77fc72304c13cb10dafc800bbc932afd319cc34442a3f76"} pod="openshift-machine-config-operator/machine-config-daemon-6dfbn" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 03 00:34:18 crc kubenswrapper[4953]: I1203 00:34:18.946600 4953 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-6dfbn" podUID="53ffb1e9-9a2e-4869-80d7-776f4aa7ef8b" containerName="machine-config-daemon" containerID="cri-o://150ffc325f215c6da77fc72304c13cb10dafc800bbc932afd319cc34442a3f76" gracePeriod=600 Dec 03 00:34:18 crc kubenswrapper[4953]: I1203 00:34:18.962596 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bwwlt\" (UniqueName: \"kubernetes.io/projected/c668590f-b40e-4c3c-9fa5-33cfcb689fc2-kube-api-access-bwwlt\") pod \"smart-gateway-operator-bundle-1-build\" (UID: \"c668590f-b40e-4c3c-9fa5-33cfcb689fc2\") " pod="service-telemetry/smart-gateway-operator-bundle-1-build" Dec 03 00:34:18 crc kubenswrapper[4953]: I1203 00:34:18.983698 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/smart-gateway-operator-bundle-1-build" Dec 03 00:34:19 crc kubenswrapper[4953]: I1203 00:34:19.211825 4953 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/smart-gateway-operator-bundle-1-build"] Dec 03 00:34:19 crc kubenswrapper[4953]: E1203 00:34:19.583584 4953 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6dfbn_openshift-machine-config-operator(53ffb1e9-9a2e-4869-80d7-776f4aa7ef8b)\"" pod="openshift-machine-config-operator/machine-config-daemon-6dfbn" podUID="53ffb1e9-9a2e-4869-80d7-776f4aa7ef8b" Dec 03 00:34:19 crc kubenswrapper[4953]: I1203 00:34:19.586132 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/smart-gateway-operator-bundle-1-build" event={"ID":"c668590f-b40e-4c3c-9fa5-33cfcb689fc2","Type":"ContainerStarted","Data":"c6561219aec20a1d2fe783d1c29fe9039b37748e18a43f060ad617826ad78c43"} Dec 03 00:34:19 crc kubenswrapper[4953]: I1203 00:34:19.586198 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/smart-gateway-operator-bundle-1-build" event={"ID":"c668590f-b40e-4c3c-9fa5-33cfcb689fc2","Type":"ContainerStarted","Data":"3dbe48756876dbe35cb7753a0fa8e7c8fc8a69e47885ef3df8ea3be144d4cdd7"} Dec 03 00:34:19 crc kubenswrapper[4953]: I1203 00:34:19.589007 4953 generic.go:334] "Generic (PLEG): container finished" podID="53ffb1e9-9a2e-4869-80d7-776f4aa7ef8b" containerID="150ffc325f215c6da77fc72304c13cb10dafc800bbc932afd319cc34442a3f76" exitCode=0 Dec 03 00:34:19 crc kubenswrapper[4953]: I1203 00:34:19.589086 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-6dfbn" event={"ID":"53ffb1e9-9a2e-4869-80d7-776f4aa7ef8b","Type":"ContainerDied","Data":"150ffc325f215c6da77fc72304c13cb10dafc800bbc932afd319cc34442a3f76"} Dec 03 00:34:19 crc kubenswrapper[4953]: I1203 00:34:19.589127 4953 scope.go:117] "RemoveContainer" containerID="995d5837fd18ae79a3033d54ed80db8a681ac98304a63c306fc50ba95ea36238" Dec 03 00:34:19 crc kubenswrapper[4953]: I1203 00:34:19.589774 4953 scope.go:117] "RemoveContainer" containerID="150ffc325f215c6da77fc72304c13cb10dafc800bbc932afd319cc34442a3f76" Dec 03 00:34:19 crc kubenswrapper[4953]: E1203 00:34:19.590049 4953 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6dfbn_openshift-machine-config-operator(53ffb1e9-9a2e-4869-80d7-776f4aa7ef8b)\"" pod="openshift-machine-config-operator/machine-config-daemon-6dfbn" podUID="53ffb1e9-9a2e-4869-80d7-776f4aa7ef8b" Dec 03 00:34:20 crc kubenswrapper[4953]: I1203 00:34:20.603622 4953 generic.go:334] "Generic (PLEG): container finished" podID="c668590f-b40e-4c3c-9fa5-33cfcb689fc2" containerID="c6561219aec20a1d2fe783d1c29fe9039b37748e18a43f060ad617826ad78c43" exitCode=0 Dec 03 00:34:20 crc kubenswrapper[4953]: I1203 00:34:20.603705 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/smart-gateway-operator-bundle-1-build" event={"ID":"c668590f-b40e-4c3c-9fa5-33cfcb689fc2","Type":"ContainerDied","Data":"c6561219aec20a1d2fe783d1c29fe9039b37748e18a43f060ad617826ad78c43"} Dec 03 00:34:20 crc kubenswrapper[4953]: I1203 00:34:20.670495 4953 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-5qxfs" Dec 03 00:34:20 crc kubenswrapper[4953]: I1203 00:34:20.671093 4953 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-5qxfs" Dec 03 00:34:20 crc kubenswrapper[4953]: I1203 00:34:20.738316 4953 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-5qxfs" Dec 03 00:34:21 crc kubenswrapper[4953]: I1203 00:34:21.613616 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/smart-gateway-operator-bundle-1-build" event={"ID":"c668590f-b40e-4c3c-9fa5-33cfcb689fc2","Type":"ContainerStarted","Data":"dd54eea3f4a21825be55d64ec038a1e721a99ec10166dd9fd1ef4a28c3b9c664"} Dec 03 00:34:21 crc kubenswrapper[4953]: I1203 00:34:21.654402 4953 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="service-telemetry/smart-gateway-operator-bundle-1-build" podStartSLOduration=3.6543584669999998 podStartE2EDuration="3.654358467s" podCreationTimestamp="2025-12-03 00:34:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 00:34:21.643838258 +0000 UTC m=+1650.628119071" watchObservedRunningTime="2025-12-03 00:34:21.654358467 +0000 UTC m=+1650.638639290" Dec 03 00:34:21 crc kubenswrapper[4953]: I1203 00:34:21.667371 4953 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-5qxfs" Dec 03 00:34:21 crc kubenswrapper[4953]: I1203 00:34:21.738562 4953 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-5qxfs"] Dec 03 00:34:22 crc kubenswrapper[4953]: I1203 00:34:22.625867 4953 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_smart-gateway-operator-bundle-1-build_c668590f-b40e-4c3c-9fa5-33cfcb689fc2/docker-build/0.log" Dec 03 00:34:22 crc kubenswrapper[4953]: I1203 00:34:22.627026 4953 generic.go:334] "Generic (PLEG): container finished" podID="c668590f-b40e-4c3c-9fa5-33cfcb689fc2" containerID="dd54eea3f4a21825be55d64ec038a1e721a99ec10166dd9fd1ef4a28c3b9c664" exitCode=1 Dec 03 00:34:22 crc kubenswrapper[4953]: I1203 00:34:22.627204 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/smart-gateway-operator-bundle-1-build" event={"ID":"c668590f-b40e-4c3c-9fa5-33cfcb689fc2","Type":"ContainerDied","Data":"dd54eea3f4a21825be55d64ec038a1e721a99ec10166dd9fd1ef4a28c3b9c664"} Dec 03 00:34:23 crc kubenswrapper[4953]: I1203 00:34:23.641078 4953 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-5qxfs" podUID="aae059a0-2ee5-4dee-9b0c-154a358dceda" containerName="registry-server" containerID="cri-o://61039323d67c2756708f022e475445f997dd5cbca40aada01fdd770bff8a3cf0" gracePeriod=2 Dec 03 00:34:23 crc kubenswrapper[4953]: I1203 00:34:23.945535 4953 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_smart-gateway-operator-bundle-1-build_c668590f-b40e-4c3c-9fa5-33cfcb689fc2/docker-build/0.log" Dec 03 00:34:23 crc kubenswrapper[4953]: I1203 00:34:23.947647 4953 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/smart-gateway-operator-bundle-1-build" Dec 03 00:34:24 crc kubenswrapper[4953]: I1203 00:34:24.072852 4953 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-5qxfs" Dec 03 00:34:24 crc kubenswrapper[4953]: I1203 00:34:24.120040 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/c668590f-b40e-4c3c-9fa5-33cfcb689fc2-buildcachedir\") pod \"c668590f-b40e-4c3c-9fa5-33cfcb689fc2\" (UID: \"c668590f-b40e-4c3c-9fa5-33cfcb689fc2\") " Dec 03 00:34:24 crc kubenswrapper[4953]: I1203 00:34:24.120201 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/c668590f-b40e-4c3c-9fa5-33cfcb689fc2-buildworkdir\") pod \"c668590f-b40e-4c3c-9fa5-33cfcb689fc2\" (UID: \"c668590f-b40e-4c3c-9fa5-33cfcb689fc2\") " Dec 03 00:34:24 crc kubenswrapper[4953]: I1203 00:34:24.120242 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c668590f-b40e-4c3c-9fa5-33cfcb689fc2-buildcachedir" (OuterVolumeSpecName: "buildcachedir") pod "c668590f-b40e-4c3c-9fa5-33cfcb689fc2" (UID: "c668590f-b40e-4c3c-9fa5-33cfcb689fc2"). InnerVolumeSpecName "buildcachedir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 00:34:24 crc kubenswrapper[4953]: I1203 00:34:24.120260 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bwwlt\" (UniqueName: \"kubernetes.io/projected/c668590f-b40e-4c3c-9fa5-33cfcb689fc2-kube-api-access-bwwlt\") pod \"c668590f-b40e-4c3c-9fa5-33cfcb689fc2\" (UID: \"c668590f-b40e-4c3c-9fa5-33cfcb689fc2\") " Dec 03 00:34:24 crc kubenswrapper[4953]: I1203 00:34:24.120306 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"builder-dockercfg-9prwd-push\" (UniqueName: \"kubernetes.io/secret/c668590f-b40e-4c3c-9fa5-33cfcb689fc2-builder-dockercfg-9prwd-push\") pod \"c668590f-b40e-4c3c-9fa5-33cfcb689fc2\" (UID: \"c668590f-b40e-4c3c-9fa5-33cfcb689fc2\") " Dec 03 00:34:24 crc kubenswrapper[4953]: I1203 00:34:24.120323 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"builder-dockercfg-9prwd-pull\" (UniqueName: \"kubernetes.io/secret/c668590f-b40e-4c3c-9fa5-33cfcb689fc2-builder-dockercfg-9prwd-pull\") pod \"c668590f-b40e-4c3c-9fa5-33cfcb689fc2\" (UID: \"c668590f-b40e-4c3c-9fa5-33cfcb689fc2\") " Dec 03 00:34:24 crc kubenswrapper[4953]: I1203 00:34:24.120366 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/c668590f-b40e-4c3c-9fa5-33cfcb689fc2-container-storage-run\") pod \"c668590f-b40e-4c3c-9fa5-33cfcb689fc2\" (UID: \"c668590f-b40e-4c3c-9fa5-33cfcb689fc2\") " Dec 03 00:34:24 crc kubenswrapper[4953]: I1203 00:34:24.120403 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/c668590f-b40e-4c3c-9fa5-33cfcb689fc2-container-storage-root\") pod \"c668590f-b40e-4c3c-9fa5-33cfcb689fc2\" (UID: \"c668590f-b40e-4c3c-9fa5-33cfcb689fc2\") " Dec 03 00:34:24 crc kubenswrapper[4953]: I1203 00:34:24.120444 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/c668590f-b40e-4c3c-9fa5-33cfcb689fc2-build-system-configs\") pod \"c668590f-b40e-4c3c-9fa5-33cfcb689fc2\" (UID: \"c668590f-b40e-4c3c-9fa5-33cfcb689fc2\") " Dec 03 00:34:24 crc kubenswrapper[4953]: I1203 00:34:24.120469 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/c668590f-b40e-4c3c-9fa5-33cfcb689fc2-build-blob-cache\") pod \"c668590f-b40e-4c3c-9fa5-33cfcb689fc2\" (UID: \"c668590f-b40e-4c3c-9fa5-33cfcb689fc2\") " Dec 03 00:34:24 crc kubenswrapper[4953]: I1203 00:34:24.120515 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/c668590f-b40e-4c3c-9fa5-33cfcb689fc2-build-ca-bundles\") pod \"c668590f-b40e-4c3c-9fa5-33cfcb689fc2\" (UID: \"c668590f-b40e-4c3c-9fa5-33cfcb689fc2\") " Dec 03 00:34:24 crc kubenswrapper[4953]: I1203 00:34:24.120559 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/c668590f-b40e-4c3c-9fa5-33cfcb689fc2-node-pullsecrets\") pod \"c668590f-b40e-4c3c-9fa5-33cfcb689fc2\" (UID: \"c668590f-b40e-4c3c-9fa5-33cfcb689fc2\") " Dec 03 00:34:24 crc kubenswrapper[4953]: I1203 00:34:24.120592 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/c668590f-b40e-4c3c-9fa5-33cfcb689fc2-build-proxy-ca-bundles\") pod \"c668590f-b40e-4c3c-9fa5-33cfcb689fc2\" (UID: \"c668590f-b40e-4c3c-9fa5-33cfcb689fc2\") " Dec 03 00:34:24 crc kubenswrapper[4953]: I1203 00:34:24.120691 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c668590f-b40e-4c3c-9fa5-33cfcb689fc2-buildworkdir" (OuterVolumeSpecName: "buildworkdir") pod "c668590f-b40e-4c3c-9fa5-33cfcb689fc2" (UID: "c668590f-b40e-4c3c-9fa5-33cfcb689fc2"). InnerVolumeSpecName "buildworkdir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 00:34:24 crc kubenswrapper[4953]: I1203 00:34:24.120922 4953 reconciler_common.go:293] "Volume detached for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/c668590f-b40e-4c3c-9fa5-33cfcb689fc2-buildcachedir\") on node \"crc\" DevicePath \"\"" Dec 03 00:34:24 crc kubenswrapper[4953]: I1203 00:34:24.120939 4953 reconciler_common.go:293] "Volume detached for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/c668590f-b40e-4c3c-9fa5-33cfcb689fc2-buildworkdir\") on node \"crc\" DevicePath \"\"" Dec 03 00:34:24 crc kubenswrapper[4953]: I1203 00:34:24.121230 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c668590f-b40e-4c3c-9fa5-33cfcb689fc2-node-pullsecrets" (OuterVolumeSpecName: "node-pullsecrets") pod "c668590f-b40e-4c3c-9fa5-33cfcb689fc2" (UID: "c668590f-b40e-4c3c-9fa5-33cfcb689fc2"). InnerVolumeSpecName "node-pullsecrets". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 00:34:24 crc kubenswrapper[4953]: I1203 00:34:24.121368 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c668590f-b40e-4c3c-9fa5-33cfcb689fc2-build-proxy-ca-bundles" (OuterVolumeSpecName: "build-proxy-ca-bundles") pod "c668590f-b40e-4c3c-9fa5-33cfcb689fc2" (UID: "c668590f-b40e-4c3c-9fa5-33cfcb689fc2"). InnerVolumeSpecName "build-proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:34:24 crc kubenswrapper[4953]: I1203 00:34:24.121899 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c668590f-b40e-4c3c-9fa5-33cfcb689fc2-container-storage-root" (OuterVolumeSpecName: "container-storage-root") pod "c668590f-b40e-4c3c-9fa5-33cfcb689fc2" (UID: "c668590f-b40e-4c3c-9fa5-33cfcb689fc2"). InnerVolumeSpecName "container-storage-root". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 00:34:24 crc kubenswrapper[4953]: I1203 00:34:24.124623 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c668590f-b40e-4c3c-9fa5-33cfcb689fc2-build-system-configs" (OuterVolumeSpecName: "build-system-configs") pod "c668590f-b40e-4c3c-9fa5-33cfcb689fc2" (UID: "c668590f-b40e-4c3c-9fa5-33cfcb689fc2"). InnerVolumeSpecName "build-system-configs". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:34:24 crc kubenswrapper[4953]: I1203 00:34:24.124927 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c668590f-b40e-4c3c-9fa5-33cfcb689fc2-build-blob-cache" (OuterVolumeSpecName: "build-blob-cache") pod "c668590f-b40e-4c3c-9fa5-33cfcb689fc2" (UID: "c668590f-b40e-4c3c-9fa5-33cfcb689fc2"). InnerVolumeSpecName "build-blob-cache". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 00:34:24 crc kubenswrapper[4953]: I1203 00:34:24.125353 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c668590f-b40e-4c3c-9fa5-33cfcb689fc2-build-ca-bundles" (OuterVolumeSpecName: "build-ca-bundles") pod "c668590f-b40e-4c3c-9fa5-33cfcb689fc2" (UID: "c668590f-b40e-4c3c-9fa5-33cfcb689fc2"). InnerVolumeSpecName "build-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:34:24 crc kubenswrapper[4953]: I1203 00:34:24.125529 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c668590f-b40e-4c3c-9fa5-33cfcb689fc2-container-storage-run" (OuterVolumeSpecName: "container-storage-run") pod "c668590f-b40e-4c3c-9fa5-33cfcb689fc2" (UID: "c668590f-b40e-4c3c-9fa5-33cfcb689fc2"). InnerVolumeSpecName "container-storage-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 00:34:24 crc kubenswrapper[4953]: I1203 00:34:24.130509 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c668590f-b40e-4c3c-9fa5-33cfcb689fc2-builder-dockercfg-9prwd-pull" (OuterVolumeSpecName: "builder-dockercfg-9prwd-pull") pod "c668590f-b40e-4c3c-9fa5-33cfcb689fc2" (UID: "c668590f-b40e-4c3c-9fa5-33cfcb689fc2"). InnerVolumeSpecName "builder-dockercfg-9prwd-pull". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:34:24 crc kubenswrapper[4953]: I1203 00:34:24.130549 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c668590f-b40e-4c3c-9fa5-33cfcb689fc2-kube-api-access-bwwlt" (OuterVolumeSpecName: "kube-api-access-bwwlt") pod "c668590f-b40e-4c3c-9fa5-33cfcb689fc2" (UID: "c668590f-b40e-4c3c-9fa5-33cfcb689fc2"). InnerVolumeSpecName "kube-api-access-bwwlt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:34:24 crc kubenswrapper[4953]: I1203 00:34:24.131079 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c668590f-b40e-4c3c-9fa5-33cfcb689fc2-builder-dockercfg-9prwd-push" (OuterVolumeSpecName: "builder-dockercfg-9prwd-push") pod "c668590f-b40e-4c3c-9fa5-33cfcb689fc2" (UID: "c668590f-b40e-4c3c-9fa5-33cfcb689fc2"). InnerVolumeSpecName "builder-dockercfg-9prwd-push". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:34:24 crc kubenswrapper[4953]: I1203 00:34:24.222783 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aae059a0-2ee5-4dee-9b0c-154a358dceda-utilities\") pod \"aae059a0-2ee5-4dee-9b0c-154a358dceda\" (UID: \"aae059a0-2ee5-4dee-9b0c-154a358dceda\") " Dec 03 00:34:24 crc kubenswrapper[4953]: I1203 00:34:24.222876 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aae059a0-2ee5-4dee-9b0c-154a358dceda-catalog-content\") pod \"aae059a0-2ee5-4dee-9b0c-154a358dceda\" (UID: \"aae059a0-2ee5-4dee-9b0c-154a358dceda\") " Dec 03 00:34:24 crc kubenswrapper[4953]: I1203 00:34:24.223053 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-846sz\" (UniqueName: \"kubernetes.io/projected/aae059a0-2ee5-4dee-9b0c-154a358dceda-kube-api-access-846sz\") pod \"aae059a0-2ee5-4dee-9b0c-154a358dceda\" (UID: \"aae059a0-2ee5-4dee-9b0c-154a358dceda\") " Dec 03 00:34:24 crc kubenswrapper[4953]: I1203 00:34:24.223415 4953 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bwwlt\" (UniqueName: \"kubernetes.io/projected/c668590f-b40e-4c3c-9fa5-33cfcb689fc2-kube-api-access-bwwlt\") on node \"crc\" DevicePath \"\"" Dec 03 00:34:24 crc kubenswrapper[4953]: I1203 00:34:24.223444 4953 reconciler_common.go:293] "Volume detached for volume \"builder-dockercfg-9prwd-push\" (UniqueName: \"kubernetes.io/secret/c668590f-b40e-4c3c-9fa5-33cfcb689fc2-builder-dockercfg-9prwd-push\") on node \"crc\" DevicePath \"\"" Dec 03 00:34:24 crc kubenswrapper[4953]: I1203 00:34:24.223460 4953 reconciler_common.go:293] "Volume detached for volume \"builder-dockercfg-9prwd-pull\" (UniqueName: \"kubernetes.io/secret/c668590f-b40e-4c3c-9fa5-33cfcb689fc2-builder-dockercfg-9prwd-pull\") on node \"crc\" DevicePath \"\"" Dec 03 00:34:24 crc kubenswrapper[4953]: I1203 00:34:24.223475 4953 reconciler_common.go:293] "Volume detached for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/c668590f-b40e-4c3c-9fa5-33cfcb689fc2-container-storage-run\") on node \"crc\" DevicePath \"\"" Dec 03 00:34:24 crc kubenswrapper[4953]: I1203 00:34:24.223487 4953 reconciler_common.go:293] "Volume detached for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/c668590f-b40e-4c3c-9fa5-33cfcb689fc2-container-storage-root\") on node \"crc\" DevicePath \"\"" Dec 03 00:34:24 crc kubenswrapper[4953]: I1203 00:34:24.223499 4953 reconciler_common.go:293] "Volume detached for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/c668590f-b40e-4c3c-9fa5-33cfcb689fc2-build-system-configs\") on node \"crc\" DevicePath \"\"" Dec 03 00:34:24 crc kubenswrapper[4953]: I1203 00:34:24.223536 4953 reconciler_common.go:293] "Volume detached for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/c668590f-b40e-4c3c-9fa5-33cfcb689fc2-build-blob-cache\") on node \"crc\" DevicePath \"\"" Dec 03 00:34:24 crc kubenswrapper[4953]: I1203 00:34:24.223551 4953 reconciler_common.go:293] "Volume detached for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/c668590f-b40e-4c3c-9fa5-33cfcb689fc2-build-ca-bundles\") on node \"crc\" DevicePath \"\"" Dec 03 00:34:24 crc kubenswrapper[4953]: I1203 00:34:24.223566 4953 reconciler_common.go:293] "Volume detached for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/c668590f-b40e-4c3c-9fa5-33cfcb689fc2-node-pullsecrets\") on node \"crc\" DevicePath \"\"" Dec 03 00:34:24 crc kubenswrapper[4953]: I1203 00:34:24.223579 4953 reconciler_common.go:293] "Volume detached for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/c668590f-b40e-4c3c-9fa5-33cfcb689fc2-build-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Dec 03 00:34:24 crc kubenswrapper[4953]: I1203 00:34:24.224666 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/aae059a0-2ee5-4dee-9b0c-154a358dceda-utilities" (OuterVolumeSpecName: "utilities") pod "aae059a0-2ee5-4dee-9b0c-154a358dceda" (UID: "aae059a0-2ee5-4dee-9b0c-154a358dceda"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 00:34:24 crc kubenswrapper[4953]: I1203 00:34:24.227912 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/aae059a0-2ee5-4dee-9b0c-154a358dceda-kube-api-access-846sz" (OuterVolumeSpecName: "kube-api-access-846sz") pod "aae059a0-2ee5-4dee-9b0c-154a358dceda" (UID: "aae059a0-2ee5-4dee-9b0c-154a358dceda"). InnerVolumeSpecName "kube-api-access-846sz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:34:24 crc kubenswrapper[4953]: I1203 00:34:24.286507 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/aae059a0-2ee5-4dee-9b0c-154a358dceda-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "aae059a0-2ee5-4dee-9b0c-154a358dceda" (UID: "aae059a0-2ee5-4dee-9b0c-154a358dceda"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 00:34:24 crc kubenswrapper[4953]: I1203 00:34:24.325159 4953 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-846sz\" (UniqueName: \"kubernetes.io/projected/aae059a0-2ee5-4dee-9b0c-154a358dceda-kube-api-access-846sz\") on node \"crc\" DevicePath \"\"" Dec 03 00:34:24 crc kubenswrapper[4953]: I1203 00:34:24.325218 4953 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aae059a0-2ee5-4dee-9b0c-154a358dceda-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 00:34:24 crc kubenswrapper[4953]: I1203 00:34:24.325229 4953 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aae059a0-2ee5-4dee-9b0c-154a358dceda-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 00:34:24 crc kubenswrapper[4953]: I1203 00:34:24.650865 4953 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_smart-gateway-operator-bundle-1-build_c668590f-b40e-4c3c-9fa5-33cfcb689fc2/docker-build/0.log" Dec 03 00:34:24 crc kubenswrapper[4953]: I1203 00:34:24.652654 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/smart-gateway-operator-bundle-1-build" event={"ID":"c668590f-b40e-4c3c-9fa5-33cfcb689fc2","Type":"ContainerDied","Data":"3dbe48756876dbe35cb7753a0fa8e7c8fc8a69e47885ef3df8ea3be144d4cdd7"} Dec 03 00:34:24 crc kubenswrapper[4953]: I1203 00:34:24.652705 4953 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3dbe48756876dbe35cb7753a0fa8e7c8fc8a69e47885ef3df8ea3be144d4cdd7" Dec 03 00:34:24 crc kubenswrapper[4953]: I1203 00:34:24.652836 4953 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/smart-gateway-operator-bundle-1-build" Dec 03 00:34:24 crc kubenswrapper[4953]: I1203 00:34:24.655246 4953 generic.go:334] "Generic (PLEG): container finished" podID="aae059a0-2ee5-4dee-9b0c-154a358dceda" containerID="61039323d67c2756708f022e475445f997dd5cbca40aada01fdd770bff8a3cf0" exitCode=0 Dec 03 00:34:24 crc kubenswrapper[4953]: I1203 00:34:24.655292 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5qxfs" event={"ID":"aae059a0-2ee5-4dee-9b0c-154a358dceda","Type":"ContainerDied","Data":"61039323d67c2756708f022e475445f997dd5cbca40aada01fdd770bff8a3cf0"} Dec 03 00:34:24 crc kubenswrapper[4953]: I1203 00:34:24.655367 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5qxfs" event={"ID":"aae059a0-2ee5-4dee-9b0c-154a358dceda","Type":"ContainerDied","Data":"b8fec301fe33a1f62600b1d85d9741e2e3c27fcb1d5490ee8057ba410ceb3066"} Dec 03 00:34:24 crc kubenswrapper[4953]: I1203 00:34:24.655392 4953 scope.go:117] "RemoveContainer" containerID="61039323d67c2756708f022e475445f997dd5cbca40aada01fdd770bff8a3cf0" Dec 03 00:34:24 crc kubenswrapper[4953]: I1203 00:34:24.655313 4953 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-5qxfs" Dec 03 00:34:24 crc kubenswrapper[4953]: I1203 00:34:24.681414 4953 scope.go:117] "RemoveContainer" containerID="d979356ce713cf58cf310e70f827e19c3cb0ef0169ddc2d50351feb628b58067" Dec 03 00:34:24 crc kubenswrapper[4953]: I1203 00:34:24.704111 4953 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-5qxfs"] Dec 03 00:34:24 crc kubenswrapper[4953]: I1203 00:34:24.710949 4953 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-5qxfs"] Dec 03 00:34:24 crc kubenswrapper[4953]: I1203 00:34:24.726773 4953 scope.go:117] "RemoveContainer" containerID="e526fbd002ca75bcecc7b726353988108b4da843e4d7be29027defec275563dc" Dec 03 00:34:24 crc kubenswrapper[4953]: I1203 00:34:24.747242 4953 scope.go:117] "RemoveContainer" containerID="61039323d67c2756708f022e475445f997dd5cbca40aada01fdd770bff8a3cf0" Dec 03 00:34:24 crc kubenswrapper[4953]: E1203 00:34:24.747852 4953 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"61039323d67c2756708f022e475445f997dd5cbca40aada01fdd770bff8a3cf0\": container with ID starting with 61039323d67c2756708f022e475445f997dd5cbca40aada01fdd770bff8a3cf0 not found: ID does not exist" containerID="61039323d67c2756708f022e475445f997dd5cbca40aada01fdd770bff8a3cf0" Dec 03 00:34:24 crc kubenswrapper[4953]: I1203 00:34:24.747885 4953 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"61039323d67c2756708f022e475445f997dd5cbca40aada01fdd770bff8a3cf0"} err="failed to get container status \"61039323d67c2756708f022e475445f997dd5cbca40aada01fdd770bff8a3cf0\": rpc error: code = NotFound desc = could not find container \"61039323d67c2756708f022e475445f997dd5cbca40aada01fdd770bff8a3cf0\": container with ID starting with 61039323d67c2756708f022e475445f997dd5cbca40aada01fdd770bff8a3cf0 not found: ID does not exist" Dec 03 00:34:24 crc kubenswrapper[4953]: I1203 00:34:24.747908 4953 scope.go:117] "RemoveContainer" containerID="d979356ce713cf58cf310e70f827e19c3cb0ef0169ddc2d50351feb628b58067" Dec 03 00:34:24 crc kubenswrapper[4953]: E1203 00:34:24.748265 4953 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d979356ce713cf58cf310e70f827e19c3cb0ef0169ddc2d50351feb628b58067\": container with ID starting with d979356ce713cf58cf310e70f827e19c3cb0ef0169ddc2d50351feb628b58067 not found: ID does not exist" containerID="d979356ce713cf58cf310e70f827e19c3cb0ef0169ddc2d50351feb628b58067" Dec 03 00:34:24 crc kubenswrapper[4953]: I1203 00:34:24.748384 4953 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d979356ce713cf58cf310e70f827e19c3cb0ef0169ddc2d50351feb628b58067"} err="failed to get container status \"d979356ce713cf58cf310e70f827e19c3cb0ef0169ddc2d50351feb628b58067\": rpc error: code = NotFound desc = could not find container \"d979356ce713cf58cf310e70f827e19c3cb0ef0169ddc2d50351feb628b58067\": container with ID starting with d979356ce713cf58cf310e70f827e19c3cb0ef0169ddc2d50351feb628b58067 not found: ID does not exist" Dec 03 00:34:24 crc kubenswrapper[4953]: I1203 00:34:24.748515 4953 scope.go:117] "RemoveContainer" containerID="e526fbd002ca75bcecc7b726353988108b4da843e4d7be29027defec275563dc" Dec 03 00:34:24 crc kubenswrapper[4953]: E1203 00:34:24.748878 4953 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e526fbd002ca75bcecc7b726353988108b4da843e4d7be29027defec275563dc\": container with ID starting with e526fbd002ca75bcecc7b726353988108b4da843e4d7be29027defec275563dc not found: ID does not exist" containerID="e526fbd002ca75bcecc7b726353988108b4da843e4d7be29027defec275563dc" Dec 03 00:34:24 crc kubenswrapper[4953]: I1203 00:34:24.749060 4953 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e526fbd002ca75bcecc7b726353988108b4da843e4d7be29027defec275563dc"} err="failed to get container status \"e526fbd002ca75bcecc7b726353988108b4da843e4d7be29027defec275563dc\": rpc error: code = NotFound desc = could not find container \"e526fbd002ca75bcecc7b726353988108b4da843e4d7be29027defec275563dc\": container with ID starting with e526fbd002ca75bcecc7b726353988108b4da843e4d7be29027defec275563dc not found: ID does not exist" Dec 03 00:34:24 crc kubenswrapper[4953]: E1203 00:34:24.753613 4953 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc668590f_b40e_4c3c_9fa5_33cfcb689fc2.slice/crio-3dbe48756876dbe35cb7753a0fa8e7c8fc8a69e47885ef3df8ea3be144d4cdd7\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podaae059a0_2ee5_4dee_9b0c_154a358dceda.slice\": RecentStats: unable to find data in memory cache]" Dec 03 00:34:25 crc kubenswrapper[4953]: I1203 00:34:25.359553 4953 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="aae059a0-2ee5-4dee-9b0c-154a358dceda" path="/var/lib/kubelet/pods/aae059a0-2ee5-4dee-9b0c-154a358dceda/volumes" Dec 03 00:34:29 crc kubenswrapper[4953]: I1203 00:34:29.143791 4953 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["service-telemetry/smart-gateway-operator-bundle-1-build"] Dec 03 00:34:29 crc kubenswrapper[4953]: I1203 00:34:29.152323 4953 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["service-telemetry/smart-gateway-operator-bundle-1-build"] Dec 03 00:34:29 crc kubenswrapper[4953]: I1203 00:34:29.357222 4953 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c668590f-b40e-4c3c-9fa5-33cfcb689fc2" path="/var/lib/kubelet/pods/c668590f-b40e-4c3c-9fa5-33cfcb689fc2/volumes" Dec 03 00:34:30 crc kubenswrapper[4953]: I1203 00:34:30.745764 4953 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["service-telemetry/smart-gateway-operator-bundle-2-build"] Dec 03 00:34:30 crc kubenswrapper[4953]: E1203 00:34:30.746134 4953 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aae059a0-2ee5-4dee-9b0c-154a358dceda" containerName="extract-utilities" Dec 03 00:34:30 crc kubenswrapper[4953]: I1203 00:34:30.746153 4953 state_mem.go:107] "Deleted CPUSet assignment" podUID="aae059a0-2ee5-4dee-9b0c-154a358dceda" containerName="extract-utilities" Dec 03 00:34:30 crc kubenswrapper[4953]: E1203 00:34:30.746175 4953 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c668590f-b40e-4c3c-9fa5-33cfcb689fc2" containerName="manage-dockerfile" Dec 03 00:34:30 crc kubenswrapper[4953]: I1203 00:34:30.746184 4953 state_mem.go:107] "Deleted CPUSet assignment" podUID="c668590f-b40e-4c3c-9fa5-33cfcb689fc2" containerName="manage-dockerfile" Dec 03 00:34:30 crc kubenswrapper[4953]: E1203 00:34:30.746200 4953 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c668590f-b40e-4c3c-9fa5-33cfcb689fc2" containerName="docker-build" Dec 03 00:34:30 crc kubenswrapper[4953]: I1203 00:34:30.746210 4953 state_mem.go:107] "Deleted CPUSet assignment" podUID="c668590f-b40e-4c3c-9fa5-33cfcb689fc2" containerName="docker-build" Dec 03 00:34:30 crc kubenswrapper[4953]: E1203 00:34:30.746227 4953 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aae059a0-2ee5-4dee-9b0c-154a358dceda" containerName="extract-content" Dec 03 00:34:30 crc kubenswrapper[4953]: I1203 00:34:30.746234 4953 state_mem.go:107] "Deleted CPUSet assignment" podUID="aae059a0-2ee5-4dee-9b0c-154a358dceda" containerName="extract-content" Dec 03 00:34:30 crc kubenswrapper[4953]: E1203 00:34:30.746252 4953 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aae059a0-2ee5-4dee-9b0c-154a358dceda" containerName="registry-server" Dec 03 00:34:30 crc kubenswrapper[4953]: I1203 00:34:30.746259 4953 state_mem.go:107] "Deleted CPUSet assignment" podUID="aae059a0-2ee5-4dee-9b0c-154a358dceda" containerName="registry-server" Dec 03 00:34:30 crc kubenswrapper[4953]: I1203 00:34:30.746413 4953 memory_manager.go:354] "RemoveStaleState removing state" podUID="aae059a0-2ee5-4dee-9b0c-154a358dceda" containerName="registry-server" Dec 03 00:34:30 crc kubenswrapper[4953]: I1203 00:34:30.746428 4953 memory_manager.go:354] "RemoveStaleState removing state" podUID="c668590f-b40e-4c3c-9fa5-33cfcb689fc2" containerName="docker-build" Dec 03 00:34:30 crc kubenswrapper[4953]: I1203 00:34:30.747567 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/smart-gateway-operator-bundle-2-build" Dec 03 00:34:30 crc kubenswrapper[4953]: I1203 00:34:30.750447 4953 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"builder-dockercfg-9prwd" Dec 03 00:34:30 crc kubenswrapper[4953]: I1203 00:34:30.750531 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"smart-gateway-operator-bundle-2-ca" Dec 03 00:34:30 crc kubenswrapper[4953]: I1203 00:34:30.751043 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"smart-gateway-operator-bundle-2-sys-config" Dec 03 00:34:30 crc kubenswrapper[4953]: I1203 00:34:30.751045 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"smart-gateway-operator-bundle-2-global-ca" Dec 03 00:34:30 crc kubenswrapper[4953]: I1203 00:34:30.775964 4953 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/smart-gateway-operator-bundle-2-build"] Dec 03 00:34:30 crc kubenswrapper[4953]: I1203 00:34:30.929164 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/1c002f52-3790-4e97-8f26-99e2128de233-node-pullsecrets\") pod \"smart-gateway-operator-bundle-2-build\" (UID: \"1c002f52-3790-4e97-8f26-99e2128de233\") " pod="service-telemetry/smart-gateway-operator-bundle-2-build" Dec 03 00:34:30 crc kubenswrapper[4953]: I1203 00:34:30.929810 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/1c002f52-3790-4e97-8f26-99e2128de233-container-storage-run\") pod \"smart-gateway-operator-bundle-2-build\" (UID: \"1c002f52-3790-4e97-8f26-99e2128de233\") " pod="service-telemetry/smart-gateway-operator-bundle-2-build" Dec 03 00:34:30 crc kubenswrapper[4953]: I1203 00:34:30.929868 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/1c002f52-3790-4e97-8f26-99e2128de233-build-blob-cache\") pod \"smart-gateway-operator-bundle-2-build\" (UID: \"1c002f52-3790-4e97-8f26-99e2128de233\") " pod="service-telemetry/smart-gateway-operator-bundle-2-build" Dec 03 00:34:30 crc kubenswrapper[4953]: I1203 00:34:30.929914 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nnbrb\" (UniqueName: \"kubernetes.io/projected/1c002f52-3790-4e97-8f26-99e2128de233-kube-api-access-nnbrb\") pod \"smart-gateway-operator-bundle-2-build\" (UID: \"1c002f52-3790-4e97-8f26-99e2128de233\") " pod="service-telemetry/smart-gateway-operator-bundle-2-build" Dec 03 00:34:30 crc kubenswrapper[4953]: I1203 00:34:30.929942 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"builder-dockercfg-9prwd-push\" (UniqueName: \"kubernetes.io/secret/1c002f52-3790-4e97-8f26-99e2128de233-builder-dockercfg-9prwd-push\") pod \"smart-gateway-operator-bundle-2-build\" (UID: \"1c002f52-3790-4e97-8f26-99e2128de233\") " pod="service-telemetry/smart-gateway-operator-bundle-2-build" Dec 03 00:34:30 crc kubenswrapper[4953]: I1203 00:34:30.930042 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/1c002f52-3790-4e97-8f26-99e2128de233-build-system-configs\") pod \"smart-gateway-operator-bundle-2-build\" (UID: \"1c002f52-3790-4e97-8f26-99e2128de233\") " pod="service-telemetry/smart-gateway-operator-bundle-2-build" Dec 03 00:34:30 crc kubenswrapper[4953]: I1203 00:34:30.930076 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/1c002f52-3790-4e97-8f26-99e2128de233-buildworkdir\") pod \"smart-gateway-operator-bundle-2-build\" (UID: \"1c002f52-3790-4e97-8f26-99e2128de233\") " pod="service-telemetry/smart-gateway-operator-bundle-2-build" Dec 03 00:34:30 crc kubenswrapper[4953]: I1203 00:34:30.930110 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/1c002f52-3790-4e97-8f26-99e2128de233-container-storage-root\") pod \"smart-gateway-operator-bundle-2-build\" (UID: \"1c002f52-3790-4e97-8f26-99e2128de233\") " pod="service-telemetry/smart-gateway-operator-bundle-2-build" Dec 03 00:34:30 crc kubenswrapper[4953]: I1203 00:34:30.930136 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/1c002f52-3790-4e97-8f26-99e2128de233-build-proxy-ca-bundles\") pod \"smart-gateway-operator-bundle-2-build\" (UID: \"1c002f52-3790-4e97-8f26-99e2128de233\") " pod="service-telemetry/smart-gateway-operator-bundle-2-build" Dec 03 00:34:30 crc kubenswrapper[4953]: I1203 00:34:30.930329 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"builder-dockercfg-9prwd-pull\" (UniqueName: \"kubernetes.io/secret/1c002f52-3790-4e97-8f26-99e2128de233-builder-dockercfg-9prwd-pull\") pod \"smart-gateway-operator-bundle-2-build\" (UID: \"1c002f52-3790-4e97-8f26-99e2128de233\") " pod="service-telemetry/smart-gateway-operator-bundle-2-build" Dec 03 00:34:30 crc kubenswrapper[4953]: I1203 00:34:30.930457 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/1c002f52-3790-4e97-8f26-99e2128de233-build-ca-bundles\") pod \"smart-gateway-operator-bundle-2-build\" (UID: \"1c002f52-3790-4e97-8f26-99e2128de233\") " pod="service-telemetry/smart-gateway-operator-bundle-2-build" Dec 03 00:34:30 crc kubenswrapper[4953]: I1203 00:34:30.930604 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/1c002f52-3790-4e97-8f26-99e2128de233-buildcachedir\") pod \"smart-gateway-operator-bundle-2-build\" (UID: \"1c002f52-3790-4e97-8f26-99e2128de233\") " pod="service-telemetry/smart-gateway-operator-bundle-2-build" Dec 03 00:34:31 crc kubenswrapper[4953]: I1203 00:34:31.032467 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nnbrb\" (UniqueName: \"kubernetes.io/projected/1c002f52-3790-4e97-8f26-99e2128de233-kube-api-access-nnbrb\") pod \"smart-gateway-operator-bundle-2-build\" (UID: \"1c002f52-3790-4e97-8f26-99e2128de233\") " pod="service-telemetry/smart-gateway-operator-bundle-2-build" Dec 03 00:34:31 crc kubenswrapper[4953]: I1203 00:34:31.032839 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"builder-dockercfg-9prwd-push\" (UniqueName: \"kubernetes.io/secret/1c002f52-3790-4e97-8f26-99e2128de233-builder-dockercfg-9prwd-push\") pod \"smart-gateway-operator-bundle-2-build\" (UID: \"1c002f52-3790-4e97-8f26-99e2128de233\") " pod="service-telemetry/smart-gateway-operator-bundle-2-build" Dec 03 00:34:31 crc kubenswrapper[4953]: I1203 00:34:31.033013 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/1c002f52-3790-4e97-8f26-99e2128de233-build-system-configs\") pod \"smart-gateway-operator-bundle-2-build\" (UID: \"1c002f52-3790-4e97-8f26-99e2128de233\") " pod="service-telemetry/smart-gateway-operator-bundle-2-build" Dec 03 00:34:31 crc kubenswrapper[4953]: I1203 00:34:31.033140 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/1c002f52-3790-4e97-8f26-99e2128de233-buildworkdir\") pod \"smart-gateway-operator-bundle-2-build\" (UID: \"1c002f52-3790-4e97-8f26-99e2128de233\") " pod="service-telemetry/smart-gateway-operator-bundle-2-build" Dec 03 00:34:31 crc kubenswrapper[4953]: I1203 00:34:31.033283 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/1c002f52-3790-4e97-8f26-99e2128de233-container-storage-root\") pod \"smart-gateway-operator-bundle-2-build\" (UID: \"1c002f52-3790-4e97-8f26-99e2128de233\") " pod="service-telemetry/smart-gateway-operator-bundle-2-build" Dec 03 00:34:31 crc kubenswrapper[4953]: I1203 00:34:31.033421 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/1c002f52-3790-4e97-8f26-99e2128de233-build-proxy-ca-bundles\") pod \"smart-gateway-operator-bundle-2-build\" (UID: \"1c002f52-3790-4e97-8f26-99e2128de233\") " pod="service-telemetry/smart-gateway-operator-bundle-2-build" Dec 03 00:34:31 crc kubenswrapper[4953]: I1203 00:34:31.033553 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"builder-dockercfg-9prwd-pull\" (UniqueName: \"kubernetes.io/secret/1c002f52-3790-4e97-8f26-99e2128de233-builder-dockercfg-9prwd-pull\") pod \"smart-gateway-operator-bundle-2-build\" (UID: \"1c002f52-3790-4e97-8f26-99e2128de233\") " pod="service-telemetry/smart-gateway-operator-bundle-2-build" Dec 03 00:34:31 crc kubenswrapper[4953]: I1203 00:34:31.033672 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/1c002f52-3790-4e97-8f26-99e2128de233-build-ca-bundles\") pod \"smart-gateway-operator-bundle-2-build\" (UID: \"1c002f52-3790-4e97-8f26-99e2128de233\") " pod="service-telemetry/smart-gateway-operator-bundle-2-build" Dec 03 00:34:31 crc kubenswrapper[4953]: I1203 00:34:31.033783 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/1c002f52-3790-4e97-8f26-99e2128de233-buildcachedir\") pod \"smart-gateway-operator-bundle-2-build\" (UID: \"1c002f52-3790-4e97-8f26-99e2128de233\") " pod="service-telemetry/smart-gateway-operator-bundle-2-build" Dec 03 00:34:31 crc kubenswrapper[4953]: I1203 00:34:31.033921 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/1c002f52-3790-4e97-8f26-99e2128de233-node-pullsecrets\") pod \"smart-gateway-operator-bundle-2-build\" (UID: \"1c002f52-3790-4e97-8f26-99e2128de233\") " pod="service-telemetry/smart-gateway-operator-bundle-2-build" Dec 03 00:34:31 crc kubenswrapper[4953]: I1203 00:34:31.034093 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/1c002f52-3790-4e97-8f26-99e2128de233-container-storage-run\") pod \"smart-gateway-operator-bundle-2-build\" (UID: \"1c002f52-3790-4e97-8f26-99e2128de233\") " pod="service-telemetry/smart-gateway-operator-bundle-2-build" Dec 03 00:34:31 crc kubenswrapper[4953]: I1203 00:34:31.034276 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/1c002f52-3790-4e97-8f26-99e2128de233-build-blob-cache\") pod \"smart-gateway-operator-bundle-2-build\" (UID: \"1c002f52-3790-4e97-8f26-99e2128de233\") " pod="service-telemetry/smart-gateway-operator-bundle-2-build" Dec 03 00:34:31 crc kubenswrapper[4953]: I1203 00:34:31.033827 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/1c002f52-3790-4e97-8f26-99e2128de233-buildcachedir\") pod \"smart-gateway-operator-bundle-2-build\" (UID: \"1c002f52-3790-4e97-8f26-99e2128de233\") " pod="service-telemetry/smart-gateway-operator-bundle-2-build" Dec 03 00:34:31 crc kubenswrapper[4953]: I1203 00:34:31.033685 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/1c002f52-3790-4e97-8f26-99e2128de233-container-storage-root\") pod \"smart-gateway-operator-bundle-2-build\" (UID: \"1c002f52-3790-4e97-8f26-99e2128de233\") " pod="service-telemetry/smart-gateway-operator-bundle-2-build" Dec 03 00:34:31 crc kubenswrapper[4953]: I1203 00:34:31.034094 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/1c002f52-3790-4e97-8f26-99e2128de233-node-pullsecrets\") pod \"smart-gateway-operator-bundle-2-build\" (UID: \"1c002f52-3790-4e97-8f26-99e2128de233\") " pod="service-telemetry/smart-gateway-operator-bundle-2-build" Dec 03 00:34:31 crc kubenswrapper[4953]: I1203 00:34:31.033788 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/1c002f52-3790-4e97-8f26-99e2128de233-build-system-configs\") pod \"smart-gateway-operator-bundle-2-build\" (UID: \"1c002f52-3790-4e97-8f26-99e2128de233\") " pod="service-telemetry/smart-gateway-operator-bundle-2-build" Dec 03 00:34:31 crc kubenswrapper[4953]: I1203 00:34:31.034285 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/1c002f52-3790-4e97-8f26-99e2128de233-build-proxy-ca-bundles\") pod \"smart-gateway-operator-bundle-2-build\" (UID: \"1c002f52-3790-4e97-8f26-99e2128de233\") " pod="service-telemetry/smart-gateway-operator-bundle-2-build" Dec 03 00:34:31 crc kubenswrapper[4953]: I1203 00:34:31.033546 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/1c002f52-3790-4e97-8f26-99e2128de233-buildworkdir\") pod \"smart-gateway-operator-bundle-2-build\" (UID: \"1c002f52-3790-4e97-8f26-99e2128de233\") " pod="service-telemetry/smart-gateway-operator-bundle-2-build" Dec 03 00:34:31 crc kubenswrapper[4953]: I1203 00:34:31.034444 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/1c002f52-3790-4e97-8f26-99e2128de233-container-storage-run\") pod \"smart-gateway-operator-bundle-2-build\" (UID: \"1c002f52-3790-4e97-8f26-99e2128de233\") " pod="service-telemetry/smart-gateway-operator-bundle-2-build" Dec 03 00:34:31 crc kubenswrapper[4953]: I1203 00:34:31.034569 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/1c002f52-3790-4e97-8f26-99e2128de233-build-blob-cache\") pod \"smart-gateway-operator-bundle-2-build\" (UID: \"1c002f52-3790-4e97-8f26-99e2128de233\") " pod="service-telemetry/smart-gateway-operator-bundle-2-build" Dec 03 00:34:31 crc kubenswrapper[4953]: I1203 00:34:31.035262 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/1c002f52-3790-4e97-8f26-99e2128de233-build-ca-bundles\") pod \"smart-gateway-operator-bundle-2-build\" (UID: \"1c002f52-3790-4e97-8f26-99e2128de233\") " pod="service-telemetry/smart-gateway-operator-bundle-2-build" Dec 03 00:34:31 crc kubenswrapper[4953]: I1203 00:34:31.040416 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"builder-dockercfg-9prwd-pull\" (UniqueName: \"kubernetes.io/secret/1c002f52-3790-4e97-8f26-99e2128de233-builder-dockercfg-9prwd-pull\") pod \"smart-gateway-operator-bundle-2-build\" (UID: \"1c002f52-3790-4e97-8f26-99e2128de233\") " pod="service-telemetry/smart-gateway-operator-bundle-2-build" Dec 03 00:34:31 crc kubenswrapper[4953]: I1203 00:34:31.041640 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"builder-dockercfg-9prwd-push\" (UniqueName: \"kubernetes.io/secret/1c002f52-3790-4e97-8f26-99e2128de233-builder-dockercfg-9prwd-push\") pod \"smart-gateway-operator-bundle-2-build\" (UID: \"1c002f52-3790-4e97-8f26-99e2128de233\") " pod="service-telemetry/smart-gateway-operator-bundle-2-build" Dec 03 00:34:31 crc kubenswrapper[4953]: I1203 00:34:31.069229 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nnbrb\" (UniqueName: \"kubernetes.io/projected/1c002f52-3790-4e97-8f26-99e2128de233-kube-api-access-nnbrb\") pod \"smart-gateway-operator-bundle-2-build\" (UID: \"1c002f52-3790-4e97-8f26-99e2128de233\") " pod="service-telemetry/smart-gateway-operator-bundle-2-build" Dec 03 00:34:31 crc kubenswrapper[4953]: I1203 00:34:31.069560 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/smart-gateway-operator-bundle-2-build" Dec 03 00:34:31 crc kubenswrapper[4953]: I1203 00:34:31.558398 4953 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/smart-gateway-operator-bundle-2-build"] Dec 03 00:34:31 crc kubenswrapper[4953]: I1203 00:34:31.722417 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/smart-gateway-operator-bundle-2-build" event={"ID":"1c002f52-3790-4e97-8f26-99e2128de233","Type":"ContainerStarted","Data":"8b6f43e8f0741cb70764dc7a1928a305c50e51089497486c45d0f5de55946b41"} Dec 03 00:34:32 crc kubenswrapper[4953]: I1203 00:34:32.731071 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/smart-gateway-operator-bundle-2-build" event={"ID":"1c002f52-3790-4e97-8f26-99e2128de233","Type":"ContainerStarted","Data":"02b4977dd4bd3319c0062ceede3789ca29891602a10a58d6ce47a1761f05b5e7"} Dec 03 00:34:33 crc kubenswrapper[4953]: I1203 00:34:33.741930 4953 generic.go:334] "Generic (PLEG): container finished" podID="1c002f52-3790-4e97-8f26-99e2128de233" containerID="02b4977dd4bd3319c0062ceede3789ca29891602a10a58d6ce47a1761f05b5e7" exitCode=0 Dec 03 00:34:33 crc kubenswrapper[4953]: I1203 00:34:33.742034 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/smart-gateway-operator-bundle-2-build" event={"ID":"1c002f52-3790-4e97-8f26-99e2128de233","Type":"ContainerDied","Data":"02b4977dd4bd3319c0062ceede3789ca29891602a10a58d6ce47a1761f05b5e7"} Dec 03 00:34:34 crc kubenswrapper[4953]: I1203 00:34:34.348008 4953 scope.go:117] "RemoveContainer" containerID="150ffc325f215c6da77fc72304c13cb10dafc800bbc932afd319cc34442a3f76" Dec 03 00:34:34 crc kubenswrapper[4953]: E1203 00:34:34.348742 4953 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6dfbn_openshift-machine-config-operator(53ffb1e9-9a2e-4869-80d7-776f4aa7ef8b)\"" pod="openshift-machine-config-operator/machine-config-daemon-6dfbn" podUID="53ffb1e9-9a2e-4869-80d7-776f4aa7ef8b" Dec 03 00:34:34 crc kubenswrapper[4953]: I1203 00:34:34.771266 4953 generic.go:334] "Generic (PLEG): container finished" podID="1c002f52-3790-4e97-8f26-99e2128de233" containerID="fd14f201ec81e13efeafc0d8746d5f0215dcfece3fb1fe9e7da50fb14a1301d3" exitCode=0 Dec 03 00:34:34 crc kubenswrapper[4953]: I1203 00:34:34.771333 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/smart-gateway-operator-bundle-2-build" event={"ID":"1c002f52-3790-4e97-8f26-99e2128de233","Type":"ContainerDied","Data":"fd14f201ec81e13efeafc0d8746d5f0215dcfece3fb1fe9e7da50fb14a1301d3"} Dec 03 00:34:34 crc kubenswrapper[4953]: I1203 00:34:34.851917 4953 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_smart-gateway-operator-bundle-2-build_1c002f52-3790-4e97-8f26-99e2128de233/manage-dockerfile/0.log" Dec 03 00:34:35 crc kubenswrapper[4953]: I1203 00:34:35.784255 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/smart-gateway-operator-bundle-2-build" event={"ID":"1c002f52-3790-4e97-8f26-99e2128de233","Type":"ContainerStarted","Data":"b94c312a00ffabc0cf23dc3a1a2890988c54034e58b8228c626e9a5d177d4c1e"} Dec 03 00:34:35 crc kubenswrapper[4953]: I1203 00:34:35.820186 4953 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="service-telemetry/smart-gateway-operator-bundle-2-build" podStartSLOduration=5.820159159 podStartE2EDuration="5.820159159s" podCreationTimestamp="2025-12-03 00:34:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 00:34:35.80961817 +0000 UTC m=+1664.793898943" watchObservedRunningTime="2025-12-03 00:34:35.820159159 +0000 UTC m=+1664.804439932" Dec 03 00:34:38 crc kubenswrapper[4953]: I1203 00:34:38.811616 4953 generic.go:334] "Generic (PLEG): container finished" podID="1c002f52-3790-4e97-8f26-99e2128de233" containerID="b94c312a00ffabc0cf23dc3a1a2890988c54034e58b8228c626e9a5d177d4c1e" exitCode=0 Dec 03 00:34:38 crc kubenswrapper[4953]: I1203 00:34:38.811693 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/smart-gateway-operator-bundle-2-build" event={"ID":"1c002f52-3790-4e97-8f26-99e2128de233","Type":"ContainerDied","Data":"b94c312a00ffabc0cf23dc3a1a2890988c54034e58b8228c626e9a5d177d4c1e"} Dec 03 00:34:40 crc kubenswrapper[4953]: I1203 00:34:40.115835 4953 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/smart-gateway-operator-bundle-2-build" Dec 03 00:34:40 crc kubenswrapper[4953]: I1203 00:34:40.299702 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/1c002f52-3790-4e97-8f26-99e2128de233-container-storage-run\") pod \"1c002f52-3790-4e97-8f26-99e2128de233\" (UID: \"1c002f52-3790-4e97-8f26-99e2128de233\") " Dec 03 00:34:40 crc kubenswrapper[4953]: I1203 00:34:40.299822 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/1c002f52-3790-4e97-8f26-99e2128de233-buildworkdir\") pod \"1c002f52-3790-4e97-8f26-99e2128de233\" (UID: \"1c002f52-3790-4e97-8f26-99e2128de233\") " Dec 03 00:34:40 crc kubenswrapper[4953]: I1203 00:34:40.299884 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/1c002f52-3790-4e97-8f26-99e2128de233-build-proxy-ca-bundles\") pod \"1c002f52-3790-4e97-8f26-99e2128de233\" (UID: \"1c002f52-3790-4e97-8f26-99e2128de233\") " Dec 03 00:34:40 crc kubenswrapper[4953]: I1203 00:34:40.299931 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/1c002f52-3790-4e97-8f26-99e2128de233-build-system-configs\") pod \"1c002f52-3790-4e97-8f26-99e2128de233\" (UID: \"1c002f52-3790-4e97-8f26-99e2128de233\") " Dec 03 00:34:40 crc kubenswrapper[4953]: I1203 00:34:40.300050 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/1c002f52-3790-4e97-8f26-99e2128de233-container-storage-root\") pod \"1c002f52-3790-4e97-8f26-99e2128de233\" (UID: \"1c002f52-3790-4e97-8f26-99e2128de233\") " Dec 03 00:34:40 crc kubenswrapper[4953]: I1203 00:34:40.300137 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/1c002f52-3790-4e97-8f26-99e2128de233-build-ca-bundles\") pod \"1c002f52-3790-4e97-8f26-99e2128de233\" (UID: \"1c002f52-3790-4e97-8f26-99e2128de233\") " Dec 03 00:34:40 crc kubenswrapper[4953]: I1203 00:34:40.300197 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/1c002f52-3790-4e97-8f26-99e2128de233-node-pullsecrets\") pod \"1c002f52-3790-4e97-8f26-99e2128de233\" (UID: \"1c002f52-3790-4e97-8f26-99e2128de233\") " Dec 03 00:34:40 crc kubenswrapper[4953]: I1203 00:34:40.300249 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/1c002f52-3790-4e97-8f26-99e2128de233-build-blob-cache\") pod \"1c002f52-3790-4e97-8f26-99e2128de233\" (UID: \"1c002f52-3790-4e97-8f26-99e2128de233\") " Dec 03 00:34:40 crc kubenswrapper[4953]: I1203 00:34:40.300342 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"builder-dockercfg-9prwd-pull\" (UniqueName: \"kubernetes.io/secret/1c002f52-3790-4e97-8f26-99e2128de233-builder-dockercfg-9prwd-pull\") pod \"1c002f52-3790-4e97-8f26-99e2128de233\" (UID: \"1c002f52-3790-4e97-8f26-99e2128de233\") " Dec 03 00:34:40 crc kubenswrapper[4953]: I1203 00:34:40.300370 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/1c002f52-3790-4e97-8f26-99e2128de233-buildcachedir\") pod \"1c002f52-3790-4e97-8f26-99e2128de233\" (UID: \"1c002f52-3790-4e97-8f26-99e2128de233\") " Dec 03 00:34:40 crc kubenswrapper[4953]: I1203 00:34:40.300402 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nnbrb\" (UniqueName: \"kubernetes.io/projected/1c002f52-3790-4e97-8f26-99e2128de233-kube-api-access-nnbrb\") pod \"1c002f52-3790-4e97-8f26-99e2128de233\" (UID: \"1c002f52-3790-4e97-8f26-99e2128de233\") " Dec 03 00:34:40 crc kubenswrapper[4953]: I1203 00:34:40.300431 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"builder-dockercfg-9prwd-push\" (UniqueName: \"kubernetes.io/secret/1c002f52-3790-4e97-8f26-99e2128de233-builder-dockercfg-9prwd-push\") pod \"1c002f52-3790-4e97-8f26-99e2128de233\" (UID: \"1c002f52-3790-4e97-8f26-99e2128de233\") " Dec 03 00:34:40 crc kubenswrapper[4953]: I1203 00:34:40.302962 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1c002f52-3790-4e97-8f26-99e2128de233-buildcachedir" (OuterVolumeSpecName: "buildcachedir") pod "1c002f52-3790-4e97-8f26-99e2128de233" (UID: "1c002f52-3790-4e97-8f26-99e2128de233"). InnerVolumeSpecName "buildcachedir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 00:34:40 crc kubenswrapper[4953]: I1203 00:34:40.303674 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1c002f52-3790-4e97-8f26-99e2128de233-node-pullsecrets" (OuterVolumeSpecName: "node-pullsecrets") pod "1c002f52-3790-4e97-8f26-99e2128de233" (UID: "1c002f52-3790-4e97-8f26-99e2128de233"). InnerVolumeSpecName "node-pullsecrets". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 00:34:40 crc kubenswrapper[4953]: I1203 00:34:40.303681 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1c002f52-3790-4e97-8f26-99e2128de233-build-proxy-ca-bundles" (OuterVolumeSpecName: "build-proxy-ca-bundles") pod "1c002f52-3790-4e97-8f26-99e2128de233" (UID: "1c002f52-3790-4e97-8f26-99e2128de233"). InnerVolumeSpecName "build-proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:34:40 crc kubenswrapper[4953]: I1203 00:34:40.303783 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1c002f52-3790-4e97-8f26-99e2128de233-build-ca-bundles" (OuterVolumeSpecName: "build-ca-bundles") pod "1c002f52-3790-4e97-8f26-99e2128de233" (UID: "1c002f52-3790-4e97-8f26-99e2128de233"). InnerVolumeSpecName "build-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:34:40 crc kubenswrapper[4953]: I1203 00:34:40.304138 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1c002f52-3790-4e97-8f26-99e2128de233-build-system-configs" (OuterVolumeSpecName: "build-system-configs") pod "1c002f52-3790-4e97-8f26-99e2128de233" (UID: "1c002f52-3790-4e97-8f26-99e2128de233"). InnerVolumeSpecName "build-system-configs". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:34:40 crc kubenswrapper[4953]: I1203 00:34:40.304411 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1c002f52-3790-4e97-8f26-99e2128de233-buildworkdir" (OuterVolumeSpecName: "buildworkdir") pod "1c002f52-3790-4e97-8f26-99e2128de233" (UID: "1c002f52-3790-4e97-8f26-99e2128de233"). InnerVolumeSpecName "buildworkdir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 00:34:40 crc kubenswrapper[4953]: I1203 00:34:40.304863 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1c002f52-3790-4e97-8f26-99e2128de233-container-storage-run" (OuterVolumeSpecName: "container-storage-run") pod "1c002f52-3790-4e97-8f26-99e2128de233" (UID: "1c002f52-3790-4e97-8f26-99e2128de233"). InnerVolumeSpecName "container-storage-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 00:34:40 crc kubenswrapper[4953]: I1203 00:34:40.309378 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1c002f52-3790-4e97-8f26-99e2128de233-build-blob-cache" (OuterVolumeSpecName: "build-blob-cache") pod "1c002f52-3790-4e97-8f26-99e2128de233" (UID: "1c002f52-3790-4e97-8f26-99e2128de233"). InnerVolumeSpecName "build-blob-cache". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 00:34:40 crc kubenswrapper[4953]: I1203 00:34:40.311199 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1c002f52-3790-4e97-8f26-99e2128de233-kube-api-access-nnbrb" (OuterVolumeSpecName: "kube-api-access-nnbrb") pod "1c002f52-3790-4e97-8f26-99e2128de233" (UID: "1c002f52-3790-4e97-8f26-99e2128de233"). InnerVolumeSpecName "kube-api-access-nnbrb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:34:40 crc kubenswrapper[4953]: I1203 00:34:40.311270 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1c002f52-3790-4e97-8f26-99e2128de233-builder-dockercfg-9prwd-push" (OuterVolumeSpecName: "builder-dockercfg-9prwd-push") pod "1c002f52-3790-4e97-8f26-99e2128de233" (UID: "1c002f52-3790-4e97-8f26-99e2128de233"). InnerVolumeSpecName "builder-dockercfg-9prwd-push". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:34:40 crc kubenswrapper[4953]: I1203 00:34:40.312317 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1c002f52-3790-4e97-8f26-99e2128de233-builder-dockercfg-9prwd-pull" (OuterVolumeSpecName: "builder-dockercfg-9prwd-pull") pod "1c002f52-3790-4e97-8f26-99e2128de233" (UID: "1c002f52-3790-4e97-8f26-99e2128de233"). InnerVolumeSpecName "builder-dockercfg-9prwd-pull". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:34:40 crc kubenswrapper[4953]: I1203 00:34:40.313700 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1c002f52-3790-4e97-8f26-99e2128de233-container-storage-root" (OuterVolumeSpecName: "container-storage-root") pod "1c002f52-3790-4e97-8f26-99e2128de233" (UID: "1c002f52-3790-4e97-8f26-99e2128de233"). InnerVolumeSpecName "container-storage-root". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 00:34:40 crc kubenswrapper[4953]: I1203 00:34:40.402275 4953 reconciler_common.go:293] "Volume detached for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/1c002f52-3790-4e97-8f26-99e2128de233-buildworkdir\") on node \"crc\" DevicePath \"\"" Dec 03 00:34:40 crc kubenswrapper[4953]: I1203 00:34:40.402344 4953 reconciler_common.go:293] "Volume detached for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/1c002f52-3790-4e97-8f26-99e2128de233-build-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Dec 03 00:34:40 crc kubenswrapper[4953]: I1203 00:34:40.402358 4953 reconciler_common.go:293] "Volume detached for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/1c002f52-3790-4e97-8f26-99e2128de233-build-system-configs\") on node \"crc\" DevicePath \"\"" Dec 03 00:34:40 crc kubenswrapper[4953]: I1203 00:34:40.402368 4953 reconciler_common.go:293] "Volume detached for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/1c002f52-3790-4e97-8f26-99e2128de233-container-storage-root\") on node \"crc\" DevicePath \"\"" Dec 03 00:34:40 crc kubenswrapper[4953]: I1203 00:34:40.402378 4953 reconciler_common.go:293] "Volume detached for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/1c002f52-3790-4e97-8f26-99e2128de233-build-ca-bundles\") on node \"crc\" DevicePath \"\"" Dec 03 00:34:40 crc kubenswrapper[4953]: I1203 00:34:40.402388 4953 reconciler_common.go:293] "Volume detached for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/1c002f52-3790-4e97-8f26-99e2128de233-node-pullsecrets\") on node \"crc\" DevicePath \"\"" Dec 03 00:34:40 crc kubenswrapper[4953]: I1203 00:34:40.402398 4953 reconciler_common.go:293] "Volume detached for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/1c002f52-3790-4e97-8f26-99e2128de233-build-blob-cache\") on node \"crc\" DevicePath \"\"" Dec 03 00:34:40 crc kubenswrapper[4953]: I1203 00:34:40.402406 4953 reconciler_common.go:293] "Volume detached for volume \"builder-dockercfg-9prwd-pull\" (UniqueName: \"kubernetes.io/secret/1c002f52-3790-4e97-8f26-99e2128de233-builder-dockercfg-9prwd-pull\") on node \"crc\" DevicePath \"\"" Dec 03 00:34:40 crc kubenswrapper[4953]: I1203 00:34:40.402416 4953 reconciler_common.go:293] "Volume detached for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/1c002f52-3790-4e97-8f26-99e2128de233-buildcachedir\") on node \"crc\" DevicePath \"\"" Dec 03 00:34:40 crc kubenswrapper[4953]: I1203 00:34:40.402425 4953 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nnbrb\" (UniqueName: \"kubernetes.io/projected/1c002f52-3790-4e97-8f26-99e2128de233-kube-api-access-nnbrb\") on node \"crc\" DevicePath \"\"" Dec 03 00:34:40 crc kubenswrapper[4953]: I1203 00:34:40.402434 4953 reconciler_common.go:293] "Volume detached for volume \"builder-dockercfg-9prwd-push\" (UniqueName: \"kubernetes.io/secret/1c002f52-3790-4e97-8f26-99e2128de233-builder-dockercfg-9prwd-push\") on node \"crc\" DevicePath \"\"" Dec 03 00:34:40 crc kubenswrapper[4953]: I1203 00:34:40.402445 4953 reconciler_common.go:293] "Volume detached for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/1c002f52-3790-4e97-8f26-99e2128de233-container-storage-run\") on node \"crc\" DevicePath \"\"" Dec 03 00:34:40 crc kubenswrapper[4953]: I1203 00:34:40.836603 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/smart-gateway-operator-bundle-2-build" event={"ID":"1c002f52-3790-4e97-8f26-99e2128de233","Type":"ContainerDied","Data":"8b6f43e8f0741cb70764dc7a1928a305c50e51089497486c45d0f5de55946b41"} Dec 03 00:34:40 crc kubenswrapper[4953]: I1203 00:34:40.836670 4953 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8b6f43e8f0741cb70764dc7a1928a305c50e51089497486c45d0f5de55946b41" Dec 03 00:34:40 crc kubenswrapper[4953]: I1203 00:34:40.836708 4953 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/smart-gateway-operator-bundle-2-build" Dec 03 00:34:47 crc kubenswrapper[4953]: I1203 00:34:47.348305 4953 scope.go:117] "RemoveContainer" containerID="150ffc325f215c6da77fc72304c13cb10dafc800bbc932afd319cc34442a3f76" Dec 03 00:34:47 crc kubenswrapper[4953]: E1203 00:34:47.349622 4953 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6dfbn_openshift-machine-config-operator(53ffb1e9-9a2e-4869-80d7-776f4aa7ef8b)\"" pod="openshift-machine-config-operator/machine-config-daemon-6dfbn" podUID="53ffb1e9-9a2e-4869-80d7-776f4aa7ef8b" Dec 03 00:34:58 crc kubenswrapper[4953]: I1203 00:34:58.059298 4953 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["service-telemetry/service-telemetry-framework-index-1-build"] Dec 03 00:34:58 crc kubenswrapper[4953]: E1203 00:34:58.062585 4953 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1c002f52-3790-4e97-8f26-99e2128de233" containerName="manage-dockerfile" Dec 03 00:34:58 crc kubenswrapper[4953]: I1203 00:34:58.062728 4953 state_mem.go:107] "Deleted CPUSet assignment" podUID="1c002f52-3790-4e97-8f26-99e2128de233" containerName="manage-dockerfile" Dec 03 00:34:58 crc kubenswrapper[4953]: E1203 00:34:58.062814 4953 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1c002f52-3790-4e97-8f26-99e2128de233" containerName="git-clone" Dec 03 00:34:58 crc kubenswrapper[4953]: I1203 00:34:58.062901 4953 state_mem.go:107] "Deleted CPUSet assignment" podUID="1c002f52-3790-4e97-8f26-99e2128de233" containerName="git-clone" Dec 03 00:34:58 crc kubenswrapper[4953]: E1203 00:34:58.063017 4953 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1c002f52-3790-4e97-8f26-99e2128de233" containerName="docker-build" Dec 03 00:34:58 crc kubenswrapper[4953]: I1203 00:34:58.063238 4953 state_mem.go:107] "Deleted CPUSet assignment" podUID="1c002f52-3790-4e97-8f26-99e2128de233" containerName="docker-build" Dec 03 00:34:58 crc kubenswrapper[4953]: I1203 00:34:58.063527 4953 memory_manager.go:354] "RemoveStaleState removing state" podUID="1c002f52-3790-4e97-8f26-99e2128de233" containerName="docker-build" Dec 03 00:34:58 crc kubenswrapper[4953]: I1203 00:34:58.064953 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/service-telemetry-framework-index-1-build" Dec 03 00:34:58 crc kubenswrapper[4953]: I1203 00:34:58.067258 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"service-telemetry-framework-index-1-sys-config" Dec 03 00:34:58 crc kubenswrapper[4953]: I1203 00:34:58.068199 4953 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"builder-dockercfg-9prwd" Dec 03 00:34:58 crc kubenswrapper[4953]: I1203 00:34:58.068502 4953 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"service-telemetry-framework-index-dockercfg" Dec 03 00:34:58 crc kubenswrapper[4953]: I1203 00:34:58.068702 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"service-telemetry-framework-index-1-ca" Dec 03 00:34:58 crc kubenswrapper[4953]: I1203 00:34:58.069428 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"service-telemetry-framework-index-1-global-ca" Dec 03 00:34:58 crc kubenswrapper[4953]: I1203 00:34:58.078541 4953 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/service-telemetry-framework-index-1-build"] Dec 03 00:34:58 crc kubenswrapper[4953]: I1203 00:34:58.084817 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/d9fe704d-0942-4735-ac13-d4a00a41ed08-node-pullsecrets\") pod \"service-telemetry-framework-index-1-build\" (UID: \"d9fe704d-0942-4735-ac13-d4a00a41ed08\") " pod="service-telemetry/service-telemetry-framework-index-1-build" Dec 03 00:34:58 crc kubenswrapper[4953]: I1203 00:34:58.084891 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/d9fe704d-0942-4735-ac13-d4a00a41ed08-container-storage-root\") pod \"service-telemetry-framework-index-1-build\" (UID: \"d9fe704d-0942-4735-ac13-d4a00a41ed08\") " pod="service-telemetry/service-telemetry-framework-index-1-build" Dec 03 00:34:58 crc kubenswrapper[4953]: I1203 00:34:58.084931 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/d9fe704d-0942-4735-ac13-d4a00a41ed08-container-storage-run\") pod \"service-telemetry-framework-index-1-build\" (UID: \"d9fe704d-0942-4735-ac13-d4a00a41ed08\") " pod="service-telemetry/service-telemetry-framework-index-1-build" Dec 03 00:34:58 crc kubenswrapper[4953]: I1203 00:34:58.084958 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/d9fe704d-0942-4735-ac13-d4a00a41ed08-buildcachedir\") pod \"service-telemetry-framework-index-1-build\" (UID: \"d9fe704d-0942-4735-ac13-d4a00a41ed08\") " pod="service-telemetry/service-telemetry-framework-index-1-build" Dec 03 00:34:58 crc kubenswrapper[4953]: I1203 00:34:58.085003 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/d9fe704d-0942-4735-ac13-d4a00a41ed08-build-blob-cache\") pod \"service-telemetry-framework-index-1-build\" (UID: \"d9fe704d-0942-4735-ac13-d4a00a41ed08\") " pod="service-telemetry/service-telemetry-framework-index-1-build" Dec 03 00:34:58 crc kubenswrapper[4953]: I1203 00:34:58.085140 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-telemetry-framework-index-dockercfg-user-build-volume\" (UniqueName: \"kubernetes.io/secret/d9fe704d-0942-4735-ac13-d4a00a41ed08-service-telemetry-framework-index-dockercfg-user-build-volume\") pod \"service-telemetry-framework-index-1-build\" (UID: \"d9fe704d-0942-4735-ac13-d4a00a41ed08\") " pod="service-telemetry/service-telemetry-framework-index-1-build" Dec 03 00:34:58 crc kubenswrapper[4953]: I1203 00:34:58.085414 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kvt48\" (UniqueName: \"kubernetes.io/projected/d9fe704d-0942-4735-ac13-d4a00a41ed08-kube-api-access-kvt48\") pod \"service-telemetry-framework-index-1-build\" (UID: \"d9fe704d-0942-4735-ac13-d4a00a41ed08\") " pod="service-telemetry/service-telemetry-framework-index-1-build" Dec 03 00:34:58 crc kubenswrapper[4953]: I1203 00:34:58.085648 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/d9fe704d-0942-4735-ac13-d4a00a41ed08-buildworkdir\") pod \"service-telemetry-framework-index-1-build\" (UID: \"d9fe704d-0942-4735-ac13-d4a00a41ed08\") " pod="service-telemetry/service-telemetry-framework-index-1-build" Dec 03 00:34:58 crc kubenswrapper[4953]: I1203 00:34:58.085685 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/d9fe704d-0942-4735-ac13-d4a00a41ed08-build-ca-bundles\") pod \"service-telemetry-framework-index-1-build\" (UID: \"d9fe704d-0942-4735-ac13-d4a00a41ed08\") " pod="service-telemetry/service-telemetry-framework-index-1-build" Dec 03 00:34:58 crc kubenswrapper[4953]: I1203 00:34:58.085744 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"builder-dockercfg-9prwd-push\" (UniqueName: \"kubernetes.io/secret/d9fe704d-0942-4735-ac13-d4a00a41ed08-builder-dockercfg-9prwd-push\") pod \"service-telemetry-framework-index-1-build\" (UID: \"d9fe704d-0942-4735-ac13-d4a00a41ed08\") " pod="service-telemetry/service-telemetry-framework-index-1-build" Dec 03 00:34:58 crc kubenswrapper[4953]: I1203 00:34:58.085787 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/d9fe704d-0942-4735-ac13-d4a00a41ed08-build-system-configs\") pod \"service-telemetry-framework-index-1-build\" (UID: \"d9fe704d-0942-4735-ac13-d4a00a41ed08\") " pod="service-telemetry/service-telemetry-framework-index-1-build" Dec 03 00:34:58 crc kubenswrapper[4953]: I1203 00:34:58.085899 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/d9fe704d-0942-4735-ac13-d4a00a41ed08-build-proxy-ca-bundles\") pod \"service-telemetry-framework-index-1-build\" (UID: \"d9fe704d-0942-4735-ac13-d4a00a41ed08\") " pod="service-telemetry/service-telemetry-framework-index-1-build" Dec 03 00:34:58 crc kubenswrapper[4953]: I1203 00:34:58.086154 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"builder-dockercfg-9prwd-pull\" (UniqueName: \"kubernetes.io/secret/d9fe704d-0942-4735-ac13-d4a00a41ed08-builder-dockercfg-9prwd-pull\") pod \"service-telemetry-framework-index-1-build\" (UID: \"d9fe704d-0942-4735-ac13-d4a00a41ed08\") " pod="service-telemetry/service-telemetry-framework-index-1-build" Dec 03 00:34:58 crc kubenswrapper[4953]: I1203 00:34:58.188532 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/d9fe704d-0942-4735-ac13-d4a00a41ed08-buildworkdir\") pod \"service-telemetry-framework-index-1-build\" (UID: \"d9fe704d-0942-4735-ac13-d4a00a41ed08\") " pod="service-telemetry/service-telemetry-framework-index-1-build" Dec 03 00:34:58 crc kubenswrapper[4953]: I1203 00:34:58.188621 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/d9fe704d-0942-4735-ac13-d4a00a41ed08-build-ca-bundles\") pod \"service-telemetry-framework-index-1-build\" (UID: \"d9fe704d-0942-4735-ac13-d4a00a41ed08\") " pod="service-telemetry/service-telemetry-framework-index-1-build" Dec 03 00:34:58 crc kubenswrapper[4953]: I1203 00:34:58.188682 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"builder-dockercfg-9prwd-push\" (UniqueName: \"kubernetes.io/secret/d9fe704d-0942-4735-ac13-d4a00a41ed08-builder-dockercfg-9prwd-push\") pod \"service-telemetry-framework-index-1-build\" (UID: \"d9fe704d-0942-4735-ac13-d4a00a41ed08\") " pod="service-telemetry/service-telemetry-framework-index-1-build" Dec 03 00:34:58 crc kubenswrapper[4953]: I1203 00:34:58.188709 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/d9fe704d-0942-4735-ac13-d4a00a41ed08-build-system-configs\") pod \"service-telemetry-framework-index-1-build\" (UID: \"d9fe704d-0942-4735-ac13-d4a00a41ed08\") " pod="service-telemetry/service-telemetry-framework-index-1-build" Dec 03 00:34:58 crc kubenswrapper[4953]: I1203 00:34:58.188769 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/d9fe704d-0942-4735-ac13-d4a00a41ed08-build-proxy-ca-bundles\") pod \"service-telemetry-framework-index-1-build\" (UID: \"d9fe704d-0942-4735-ac13-d4a00a41ed08\") " pod="service-telemetry/service-telemetry-framework-index-1-build" Dec 03 00:34:58 crc kubenswrapper[4953]: I1203 00:34:58.188855 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"builder-dockercfg-9prwd-pull\" (UniqueName: \"kubernetes.io/secret/d9fe704d-0942-4735-ac13-d4a00a41ed08-builder-dockercfg-9prwd-pull\") pod \"service-telemetry-framework-index-1-build\" (UID: \"d9fe704d-0942-4735-ac13-d4a00a41ed08\") " pod="service-telemetry/service-telemetry-framework-index-1-build" Dec 03 00:34:58 crc kubenswrapper[4953]: I1203 00:34:58.188908 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/d9fe704d-0942-4735-ac13-d4a00a41ed08-node-pullsecrets\") pod \"service-telemetry-framework-index-1-build\" (UID: \"d9fe704d-0942-4735-ac13-d4a00a41ed08\") " pod="service-telemetry/service-telemetry-framework-index-1-build" Dec 03 00:34:58 crc kubenswrapper[4953]: I1203 00:34:58.188940 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/d9fe704d-0942-4735-ac13-d4a00a41ed08-container-storage-root\") pod \"service-telemetry-framework-index-1-build\" (UID: \"d9fe704d-0942-4735-ac13-d4a00a41ed08\") " pod="service-telemetry/service-telemetry-framework-index-1-build" Dec 03 00:34:58 crc kubenswrapper[4953]: I1203 00:34:58.189000 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/d9fe704d-0942-4735-ac13-d4a00a41ed08-container-storage-run\") pod \"service-telemetry-framework-index-1-build\" (UID: \"d9fe704d-0942-4735-ac13-d4a00a41ed08\") " pod="service-telemetry/service-telemetry-framework-index-1-build" Dec 03 00:34:58 crc kubenswrapper[4953]: I1203 00:34:58.189031 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/d9fe704d-0942-4735-ac13-d4a00a41ed08-buildcachedir\") pod \"service-telemetry-framework-index-1-build\" (UID: \"d9fe704d-0942-4735-ac13-d4a00a41ed08\") " pod="service-telemetry/service-telemetry-framework-index-1-build" Dec 03 00:34:58 crc kubenswrapper[4953]: I1203 00:34:58.189055 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/d9fe704d-0942-4735-ac13-d4a00a41ed08-build-blob-cache\") pod \"service-telemetry-framework-index-1-build\" (UID: \"d9fe704d-0942-4735-ac13-d4a00a41ed08\") " pod="service-telemetry/service-telemetry-framework-index-1-build" Dec 03 00:34:58 crc kubenswrapper[4953]: I1203 00:34:58.189122 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-telemetry-framework-index-dockercfg-user-build-volume\" (UniqueName: \"kubernetes.io/secret/d9fe704d-0942-4735-ac13-d4a00a41ed08-service-telemetry-framework-index-dockercfg-user-build-volume\") pod \"service-telemetry-framework-index-1-build\" (UID: \"d9fe704d-0942-4735-ac13-d4a00a41ed08\") " pod="service-telemetry/service-telemetry-framework-index-1-build" Dec 03 00:34:58 crc kubenswrapper[4953]: I1203 00:34:58.189177 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kvt48\" (UniqueName: \"kubernetes.io/projected/d9fe704d-0942-4735-ac13-d4a00a41ed08-kube-api-access-kvt48\") pod \"service-telemetry-framework-index-1-build\" (UID: \"d9fe704d-0942-4735-ac13-d4a00a41ed08\") " pod="service-telemetry/service-telemetry-framework-index-1-build" Dec 03 00:34:58 crc kubenswrapper[4953]: I1203 00:34:58.189286 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/d9fe704d-0942-4735-ac13-d4a00a41ed08-buildworkdir\") pod \"service-telemetry-framework-index-1-build\" (UID: \"d9fe704d-0942-4735-ac13-d4a00a41ed08\") " pod="service-telemetry/service-telemetry-framework-index-1-build" Dec 03 00:34:58 crc kubenswrapper[4953]: I1203 00:34:58.189526 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/d9fe704d-0942-4735-ac13-d4a00a41ed08-buildcachedir\") pod \"service-telemetry-framework-index-1-build\" (UID: \"d9fe704d-0942-4735-ac13-d4a00a41ed08\") " pod="service-telemetry/service-telemetry-framework-index-1-build" Dec 03 00:34:58 crc kubenswrapper[4953]: I1203 00:34:58.189583 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/d9fe704d-0942-4735-ac13-d4a00a41ed08-container-storage-root\") pod \"service-telemetry-framework-index-1-build\" (UID: \"d9fe704d-0942-4735-ac13-d4a00a41ed08\") " pod="service-telemetry/service-telemetry-framework-index-1-build" Dec 03 00:34:58 crc kubenswrapper[4953]: I1203 00:34:58.189874 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/d9fe704d-0942-4735-ac13-d4a00a41ed08-build-ca-bundles\") pod \"service-telemetry-framework-index-1-build\" (UID: \"d9fe704d-0942-4735-ac13-d4a00a41ed08\") " pod="service-telemetry/service-telemetry-framework-index-1-build" Dec 03 00:34:58 crc kubenswrapper[4953]: I1203 00:34:58.189884 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/d9fe704d-0942-4735-ac13-d4a00a41ed08-container-storage-run\") pod \"service-telemetry-framework-index-1-build\" (UID: \"d9fe704d-0942-4735-ac13-d4a00a41ed08\") " pod="service-telemetry/service-telemetry-framework-index-1-build" Dec 03 00:34:58 crc kubenswrapper[4953]: I1203 00:34:58.189921 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/d9fe704d-0942-4735-ac13-d4a00a41ed08-node-pullsecrets\") pod \"service-telemetry-framework-index-1-build\" (UID: \"d9fe704d-0942-4735-ac13-d4a00a41ed08\") " pod="service-telemetry/service-telemetry-framework-index-1-build" Dec 03 00:34:58 crc kubenswrapper[4953]: I1203 00:34:58.190255 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/d9fe704d-0942-4735-ac13-d4a00a41ed08-build-blob-cache\") pod \"service-telemetry-framework-index-1-build\" (UID: \"d9fe704d-0942-4735-ac13-d4a00a41ed08\") " pod="service-telemetry/service-telemetry-framework-index-1-build" Dec 03 00:34:58 crc kubenswrapper[4953]: I1203 00:34:58.190415 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/d9fe704d-0942-4735-ac13-d4a00a41ed08-build-system-configs\") pod \"service-telemetry-framework-index-1-build\" (UID: \"d9fe704d-0942-4735-ac13-d4a00a41ed08\") " pod="service-telemetry/service-telemetry-framework-index-1-build" Dec 03 00:34:58 crc kubenswrapper[4953]: I1203 00:34:58.191022 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/d9fe704d-0942-4735-ac13-d4a00a41ed08-build-proxy-ca-bundles\") pod \"service-telemetry-framework-index-1-build\" (UID: \"d9fe704d-0942-4735-ac13-d4a00a41ed08\") " pod="service-telemetry/service-telemetry-framework-index-1-build" Dec 03 00:34:58 crc kubenswrapper[4953]: I1203 00:34:58.196213 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"builder-dockercfg-9prwd-push\" (UniqueName: \"kubernetes.io/secret/d9fe704d-0942-4735-ac13-d4a00a41ed08-builder-dockercfg-9prwd-push\") pod \"service-telemetry-framework-index-1-build\" (UID: \"d9fe704d-0942-4735-ac13-d4a00a41ed08\") " pod="service-telemetry/service-telemetry-framework-index-1-build" Dec 03 00:34:58 crc kubenswrapper[4953]: I1203 00:34:58.196555 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-telemetry-framework-index-dockercfg-user-build-volume\" (UniqueName: \"kubernetes.io/secret/d9fe704d-0942-4735-ac13-d4a00a41ed08-service-telemetry-framework-index-dockercfg-user-build-volume\") pod \"service-telemetry-framework-index-1-build\" (UID: \"d9fe704d-0942-4735-ac13-d4a00a41ed08\") " pod="service-telemetry/service-telemetry-framework-index-1-build" Dec 03 00:34:58 crc kubenswrapper[4953]: I1203 00:34:58.197467 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"builder-dockercfg-9prwd-pull\" (UniqueName: \"kubernetes.io/secret/d9fe704d-0942-4735-ac13-d4a00a41ed08-builder-dockercfg-9prwd-pull\") pod \"service-telemetry-framework-index-1-build\" (UID: \"d9fe704d-0942-4735-ac13-d4a00a41ed08\") " pod="service-telemetry/service-telemetry-framework-index-1-build" Dec 03 00:34:58 crc kubenswrapper[4953]: I1203 00:34:58.209457 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kvt48\" (UniqueName: \"kubernetes.io/projected/d9fe704d-0942-4735-ac13-d4a00a41ed08-kube-api-access-kvt48\") pod \"service-telemetry-framework-index-1-build\" (UID: \"d9fe704d-0942-4735-ac13-d4a00a41ed08\") " pod="service-telemetry/service-telemetry-framework-index-1-build" Dec 03 00:34:58 crc kubenswrapper[4953]: I1203 00:34:58.392163 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/service-telemetry-framework-index-1-build" Dec 03 00:34:58 crc kubenswrapper[4953]: I1203 00:34:58.820556 4953 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/service-telemetry-framework-index-1-build"] Dec 03 00:34:58 crc kubenswrapper[4953]: I1203 00:34:58.989619 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/service-telemetry-framework-index-1-build" event={"ID":"d9fe704d-0942-4735-ac13-d4a00a41ed08","Type":"ContainerStarted","Data":"60f494793015bd67b75d60d9080131bc465a708a57502b23672c4c8cdac973d8"} Dec 03 00:34:59 crc kubenswrapper[4953]: I1203 00:34:59.998461 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/service-telemetry-framework-index-1-build" event={"ID":"d9fe704d-0942-4735-ac13-d4a00a41ed08","Type":"ContainerStarted","Data":"9c54779a7684dd2c2da9df651d91093e1291563d0fa6db25e3f51a63658ab602"} Dec 03 00:35:01 crc kubenswrapper[4953]: I1203 00:35:01.008537 4953 generic.go:334] "Generic (PLEG): container finished" podID="d9fe704d-0942-4735-ac13-d4a00a41ed08" containerID="9c54779a7684dd2c2da9df651d91093e1291563d0fa6db25e3f51a63658ab602" exitCode=0 Dec 03 00:35:01 crc kubenswrapper[4953]: I1203 00:35:01.008607 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/service-telemetry-framework-index-1-build" event={"ID":"d9fe704d-0942-4735-ac13-d4a00a41ed08","Type":"ContainerDied","Data":"9c54779a7684dd2c2da9df651d91093e1291563d0fa6db25e3f51a63658ab602"} Dec 03 00:35:02 crc kubenswrapper[4953]: I1203 00:35:02.024602 4953 generic.go:334] "Generic (PLEG): container finished" podID="d9fe704d-0942-4735-ac13-d4a00a41ed08" containerID="f6ec2418e3ab6ff1524538a691e065a3c8a7dc1a003f1fd3f28fc4bbbb020075" exitCode=0 Dec 03 00:35:02 crc kubenswrapper[4953]: I1203 00:35:02.024853 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/service-telemetry-framework-index-1-build" event={"ID":"d9fe704d-0942-4735-ac13-d4a00a41ed08","Type":"ContainerDied","Data":"f6ec2418e3ab6ff1524538a691e065a3c8a7dc1a003f1fd3f28fc4bbbb020075"} Dec 03 00:35:02 crc kubenswrapper[4953]: I1203 00:35:02.067467 4953 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_service-telemetry-framework-index-1-build_d9fe704d-0942-4735-ac13-d4a00a41ed08/manage-dockerfile/0.log" Dec 03 00:35:02 crc kubenswrapper[4953]: I1203 00:35:02.348483 4953 scope.go:117] "RemoveContainer" containerID="150ffc325f215c6da77fc72304c13cb10dafc800bbc932afd319cc34442a3f76" Dec 03 00:35:02 crc kubenswrapper[4953]: E1203 00:35:02.349552 4953 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6dfbn_openshift-machine-config-operator(53ffb1e9-9a2e-4869-80d7-776f4aa7ef8b)\"" pod="openshift-machine-config-operator/machine-config-daemon-6dfbn" podUID="53ffb1e9-9a2e-4869-80d7-776f4aa7ef8b" Dec 03 00:35:03 crc kubenswrapper[4953]: I1203 00:35:03.036402 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/service-telemetry-framework-index-1-build" event={"ID":"d9fe704d-0942-4735-ac13-d4a00a41ed08","Type":"ContainerStarted","Data":"35257f63670afe3a8d5d8d6cb5d2550df6af73a52edc93f1de66eb89a6b3c9bb"} Dec 03 00:35:03 crc kubenswrapper[4953]: I1203 00:35:03.067400 4953 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="service-telemetry/service-telemetry-framework-index-1-build" podStartSLOduration=5.067367692 podStartE2EDuration="5.067367692s" podCreationTimestamp="2025-12-03 00:34:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 00:35:03.063684922 +0000 UTC m=+1692.047965705" watchObservedRunningTime="2025-12-03 00:35:03.067367692 +0000 UTC m=+1692.051648475" Dec 03 00:35:16 crc kubenswrapper[4953]: I1203 00:35:16.347836 4953 scope.go:117] "RemoveContainer" containerID="150ffc325f215c6da77fc72304c13cb10dafc800bbc932afd319cc34442a3f76" Dec 03 00:35:16 crc kubenswrapper[4953]: E1203 00:35:16.349144 4953 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6dfbn_openshift-machine-config-operator(53ffb1e9-9a2e-4869-80d7-776f4aa7ef8b)\"" pod="openshift-machine-config-operator/machine-config-daemon-6dfbn" podUID="53ffb1e9-9a2e-4869-80d7-776f4aa7ef8b" Dec 03 00:35:28 crc kubenswrapper[4953]: I1203 00:35:28.348472 4953 scope.go:117] "RemoveContainer" containerID="150ffc325f215c6da77fc72304c13cb10dafc800bbc932afd319cc34442a3f76" Dec 03 00:35:28 crc kubenswrapper[4953]: E1203 00:35:28.349930 4953 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6dfbn_openshift-machine-config-operator(53ffb1e9-9a2e-4869-80d7-776f4aa7ef8b)\"" pod="openshift-machine-config-operator/machine-config-daemon-6dfbn" podUID="53ffb1e9-9a2e-4869-80d7-776f4aa7ef8b" Dec 03 00:35:38 crc kubenswrapper[4953]: I1203 00:35:38.353667 4953 generic.go:334] "Generic (PLEG): container finished" podID="d9fe704d-0942-4735-ac13-d4a00a41ed08" containerID="35257f63670afe3a8d5d8d6cb5d2550df6af73a52edc93f1de66eb89a6b3c9bb" exitCode=0 Dec 03 00:35:38 crc kubenswrapper[4953]: I1203 00:35:38.353695 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/service-telemetry-framework-index-1-build" event={"ID":"d9fe704d-0942-4735-ac13-d4a00a41ed08","Type":"ContainerDied","Data":"35257f63670afe3a8d5d8d6cb5d2550df6af73a52edc93f1de66eb89a6b3c9bb"} Dec 03 00:35:39 crc kubenswrapper[4953]: I1203 00:35:39.348583 4953 scope.go:117] "RemoveContainer" containerID="150ffc325f215c6da77fc72304c13cb10dafc800bbc932afd319cc34442a3f76" Dec 03 00:35:39 crc kubenswrapper[4953]: E1203 00:35:39.349433 4953 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6dfbn_openshift-machine-config-operator(53ffb1e9-9a2e-4869-80d7-776f4aa7ef8b)\"" pod="openshift-machine-config-operator/machine-config-daemon-6dfbn" podUID="53ffb1e9-9a2e-4869-80d7-776f4aa7ef8b" Dec 03 00:35:39 crc kubenswrapper[4953]: I1203 00:35:39.675604 4953 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/service-telemetry-framework-index-1-build" Dec 03 00:35:39 crc kubenswrapper[4953]: I1203 00:35:39.756197 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/d9fe704d-0942-4735-ac13-d4a00a41ed08-build-ca-bundles\") pod \"d9fe704d-0942-4735-ac13-d4a00a41ed08\" (UID: \"d9fe704d-0942-4735-ac13-d4a00a41ed08\") " Dec 03 00:35:39 crc kubenswrapper[4953]: I1203 00:35:39.756290 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/d9fe704d-0942-4735-ac13-d4a00a41ed08-buildworkdir\") pod \"d9fe704d-0942-4735-ac13-d4a00a41ed08\" (UID: \"d9fe704d-0942-4735-ac13-d4a00a41ed08\") " Dec 03 00:35:39 crc kubenswrapper[4953]: I1203 00:35:39.756317 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/d9fe704d-0942-4735-ac13-d4a00a41ed08-container-storage-root\") pod \"d9fe704d-0942-4735-ac13-d4a00a41ed08\" (UID: \"d9fe704d-0942-4735-ac13-d4a00a41ed08\") " Dec 03 00:35:39 crc kubenswrapper[4953]: I1203 00:35:39.756354 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-telemetry-framework-index-dockercfg-user-build-volume\" (UniqueName: \"kubernetes.io/secret/d9fe704d-0942-4735-ac13-d4a00a41ed08-service-telemetry-framework-index-dockercfg-user-build-volume\") pod \"d9fe704d-0942-4735-ac13-d4a00a41ed08\" (UID: \"d9fe704d-0942-4735-ac13-d4a00a41ed08\") " Dec 03 00:35:39 crc kubenswrapper[4953]: I1203 00:35:39.756385 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/d9fe704d-0942-4735-ac13-d4a00a41ed08-build-blob-cache\") pod \"d9fe704d-0942-4735-ac13-d4a00a41ed08\" (UID: \"d9fe704d-0942-4735-ac13-d4a00a41ed08\") " Dec 03 00:35:39 crc kubenswrapper[4953]: I1203 00:35:39.756461 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kvt48\" (UniqueName: \"kubernetes.io/projected/d9fe704d-0942-4735-ac13-d4a00a41ed08-kube-api-access-kvt48\") pod \"d9fe704d-0942-4735-ac13-d4a00a41ed08\" (UID: \"d9fe704d-0942-4735-ac13-d4a00a41ed08\") " Dec 03 00:35:39 crc kubenswrapper[4953]: I1203 00:35:39.756495 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/d9fe704d-0942-4735-ac13-d4a00a41ed08-build-proxy-ca-bundles\") pod \"d9fe704d-0942-4735-ac13-d4a00a41ed08\" (UID: \"d9fe704d-0942-4735-ac13-d4a00a41ed08\") " Dec 03 00:35:39 crc kubenswrapper[4953]: I1203 00:35:39.756569 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/d9fe704d-0942-4735-ac13-d4a00a41ed08-container-storage-run\") pod \"d9fe704d-0942-4735-ac13-d4a00a41ed08\" (UID: \"d9fe704d-0942-4735-ac13-d4a00a41ed08\") " Dec 03 00:35:39 crc kubenswrapper[4953]: I1203 00:35:39.756623 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/d9fe704d-0942-4735-ac13-d4a00a41ed08-node-pullsecrets\") pod \"d9fe704d-0942-4735-ac13-d4a00a41ed08\" (UID: \"d9fe704d-0942-4735-ac13-d4a00a41ed08\") " Dec 03 00:35:39 crc kubenswrapper[4953]: I1203 00:35:39.756645 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/d9fe704d-0942-4735-ac13-d4a00a41ed08-build-system-configs\") pod \"d9fe704d-0942-4735-ac13-d4a00a41ed08\" (UID: \"d9fe704d-0942-4735-ac13-d4a00a41ed08\") " Dec 03 00:35:39 crc kubenswrapper[4953]: I1203 00:35:39.756730 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"builder-dockercfg-9prwd-push\" (UniqueName: \"kubernetes.io/secret/d9fe704d-0942-4735-ac13-d4a00a41ed08-builder-dockercfg-9prwd-push\") pod \"d9fe704d-0942-4735-ac13-d4a00a41ed08\" (UID: \"d9fe704d-0942-4735-ac13-d4a00a41ed08\") " Dec 03 00:35:39 crc kubenswrapper[4953]: I1203 00:35:39.756755 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"builder-dockercfg-9prwd-pull\" (UniqueName: \"kubernetes.io/secret/d9fe704d-0942-4735-ac13-d4a00a41ed08-builder-dockercfg-9prwd-pull\") pod \"d9fe704d-0942-4735-ac13-d4a00a41ed08\" (UID: \"d9fe704d-0942-4735-ac13-d4a00a41ed08\") " Dec 03 00:35:39 crc kubenswrapper[4953]: I1203 00:35:39.756804 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/d9fe704d-0942-4735-ac13-d4a00a41ed08-buildcachedir\") pod \"d9fe704d-0942-4735-ac13-d4a00a41ed08\" (UID: \"d9fe704d-0942-4735-ac13-d4a00a41ed08\") " Dec 03 00:35:39 crc kubenswrapper[4953]: I1203 00:35:39.757147 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d9fe704d-0942-4735-ac13-d4a00a41ed08-buildworkdir" (OuterVolumeSpecName: "buildworkdir") pod "d9fe704d-0942-4735-ac13-d4a00a41ed08" (UID: "d9fe704d-0942-4735-ac13-d4a00a41ed08"). InnerVolumeSpecName "buildworkdir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 00:35:39 crc kubenswrapper[4953]: I1203 00:35:39.757250 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d9fe704d-0942-4735-ac13-d4a00a41ed08-buildcachedir" (OuterVolumeSpecName: "buildcachedir") pod "d9fe704d-0942-4735-ac13-d4a00a41ed08" (UID: "d9fe704d-0942-4735-ac13-d4a00a41ed08"). InnerVolumeSpecName "buildcachedir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 00:35:39 crc kubenswrapper[4953]: I1203 00:35:39.757366 4953 reconciler_common.go:293] "Volume detached for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/d9fe704d-0942-4735-ac13-d4a00a41ed08-buildworkdir\") on node \"crc\" DevicePath \"\"" Dec 03 00:35:39 crc kubenswrapper[4953]: I1203 00:35:39.757408 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d9fe704d-0942-4735-ac13-d4a00a41ed08-node-pullsecrets" (OuterVolumeSpecName: "node-pullsecrets") pod "d9fe704d-0942-4735-ac13-d4a00a41ed08" (UID: "d9fe704d-0942-4735-ac13-d4a00a41ed08"). InnerVolumeSpecName "node-pullsecrets". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 00:35:39 crc kubenswrapper[4953]: I1203 00:35:39.757585 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d9fe704d-0942-4735-ac13-d4a00a41ed08-build-ca-bundles" (OuterVolumeSpecName: "build-ca-bundles") pod "d9fe704d-0942-4735-ac13-d4a00a41ed08" (UID: "d9fe704d-0942-4735-ac13-d4a00a41ed08"). InnerVolumeSpecName "build-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:35:39 crc kubenswrapper[4953]: I1203 00:35:39.758020 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d9fe704d-0942-4735-ac13-d4a00a41ed08-build-proxy-ca-bundles" (OuterVolumeSpecName: "build-proxy-ca-bundles") pod "d9fe704d-0942-4735-ac13-d4a00a41ed08" (UID: "d9fe704d-0942-4735-ac13-d4a00a41ed08"). InnerVolumeSpecName "build-proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:35:39 crc kubenswrapper[4953]: I1203 00:35:39.758689 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d9fe704d-0942-4735-ac13-d4a00a41ed08-build-system-configs" (OuterVolumeSpecName: "build-system-configs") pod "d9fe704d-0942-4735-ac13-d4a00a41ed08" (UID: "d9fe704d-0942-4735-ac13-d4a00a41ed08"). InnerVolumeSpecName "build-system-configs". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:35:39 crc kubenswrapper[4953]: I1203 00:35:39.769099 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d9fe704d-0942-4735-ac13-d4a00a41ed08-service-telemetry-framework-index-dockercfg-user-build-volume" (OuterVolumeSpecName: "service-telemetry-framework-index-dockercfg-user-build-volume") pod "d9fe704d-0942-4735-ac13-d4a00a41ed08" (UID: "d9fe704d-0942-4735-ac13-d4a00a41ed08"). InnerVolumeSpecName "service-telemetry-framework-index-dockercfg-user-build-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:35:39 crc kubenswrapper[4953]: I1203 00:35:39.769235 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d9fe704d-0942-4735-ac13-d4a00a41ed08-builder-dockercfg-9prwd-push" (OuterVolumeSpecName: "builder-dockercfg-9prwd-push") pod "d9fe704d-0942-4735-ac13-d4a00a41ed08" (UID: "d9fe704d-0942-4735-ac13-d4a00a41ed08"). InnerVolumeSpecName "builder-dockercfg-9prwd-push". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:35:39 crc kubenswrapper[4953]: I1203 00:35:39.769142 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d9fe704d-0942-4735-ac13-d4a00a41ed08-builder-dockercfg-9prwd-pull" (OuterVolumeSpecName: "builder-dockercfg-9prwd-pull") pod "d9fe704d-0942-4735-ac13-d4a00a41ed08" (UID: "d9fe704d-0942-4735-ac13-d4a00a41ed08"). InnerVolumeSpecName "builder-dockercfg-9prwd-pull". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:35:39 crc kubenswrapper[4953]: I1203 00:35:39.771315 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d9fe704d-0942-4735-ac13-d4a00a41ed08-kube-api-access-kvt48" (OuterVolumeSpecName: "kube-api-access-kvt48") pod "d9fe704d-0942-4735-ac13-d4a00a41ed08" (UID: "d9fe704d-0942-4735-ac13-d4a00a41ed08"). InnerVolumeSpecName "kube-api-access-kvt48". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:35:39 crc kubenswrapper[4953]: I1203 00:35:39.858913 4953 reconciler_common.go:293] "Volume detached for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/d9fe704d-0942-4735-ac13-d4a00a41ed08-node-pullsecrets\") on node \"crc\" DevicePath \"\"" Dec 03 00:35:39 crc kubenswrapper[4953]: I1203 00:35:39.859544 4953 reconciler_common.go:293] "Volume detached for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/d9fe704d-0942-4735-ac13-d4a00a41ed08-build-system-configs\") on node \"crc\" DevicePath \"\"" Dec 03 00:35:39 crc kubenswrapper[4953]: I1203 00:35:39.859681 4953 reconciler_common.go:293] "Volume detached for volume \"builder-dockercfg-9prwd-push\" (UniqueName: \"kubernetes.io/secret/d9fe704d-0942-4735-ac13-d4a00a41ed08-builder-dockercfg-9prwd-push\") on node \"crc\" DevicePath \"\"" Dec 03 00:35:39 crc kubenswrapper[4953]: I1203 00:35:39.859744 4953 reconciler_common.go:293] "Volume detached for volume \"builder-dockercfg-9prwd-pull\" (UniqueName: \"kubernetes.io/secret/d9fe704d-0942-4735-ac13-d4a00a41ed08-builder-dockercfg-9prwd-pull\") on node \"crc\" DevicePath \"\"" Dec 03 00:35:39 crc kubenswrapper[4953]: I1203 00:35:39.859838 4953 reconciler_common.go:293] "Volume detached for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/d9fe704d-0942-4735-ac13-d4a00a41ed08-buildcachedir\") on node \"crc\" DevicePath \"\"" Dec 03 00:35:39 crc kubenswrapper[4953]: I1203 00:35:39.859938 4953 reconciler_common.go:293] "Volume detached for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/d9fe704d-0942-4735-ac13-d4a00a41ed08-build-ca-bundles\") on node \"crc\" DevicePath \"\"" Dec 03 00:35:39 crc kubenswrapper[4953]: I1203 00:35:39.860048 4953 reconciler_common.go:293] "Volume detached for volume \"service-telemetry-framework-index-dockercfg-user-build-volume\" (UniqueName: \"kubernetes.io/secret/d9fe704d-0942-4735-ac13-d4a00a41ed08-service-telemetry-framework-index-dockercfg-user-build-volume\") on node \"crc\" DevicePath \"\"" Dec 03 00:35:39 crc kubenswrapper[4953]: I1203 00:35:39.860141 4953 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kvt48\" (UniqueName: \"kubernetes.io/projected/d9fe704d-0942-4735-ac13-d4a00a41ed08-kube-api-access-kvt48\") on node \"crc\" DevicePath \"\"" Dec 03 00:35:39 crc kubenswrapper[4953]: I1203 00:35:39.860226 4953 reconciler_common.go:293] "Volume detached for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/d9fe704d-0942-4735-ac13-d4a00a41ed08-build-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Dec 03 00:35:40 crc kubenswrapper[4953]: I1203 00:35:40.021492 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d9fe704d-0942-4735-ac13-d4a00a41ed08-build-blob-cache" (OuterVolumeSpecName: "build-blob-cache") pod "d9fe704d-0942-4735-ac13-d4a00a41ed08" (UID: "d9fe704d-0942-4735-ac13-d4a00a41ed08"). InnerVolumeSpecName "build-blob-cache". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 00:35:40 crc kubenswrapper[4953]: I1203 00:35:40.064292 4953 reconciler_common.go:293] "Volume detached for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/d9fe704d-0942-4735-ac13-d4a00a41ed08-build-blob-cache\") on node \"crc\" DevicePath \"\"" Dec 03 00:35:40 crc kubenswrapper[4953]: I1203 00:35:40.246397 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d9fe704d-0942-4735-ac13-d4a00a41ed08-container-storage-run" (OuterVolumeSpecName: "container-storage-run") pod "d9fe704d-0942-4735-ac13-d4a00a41ed08" (UID: "d9fe704d-0942-4735-ac13-d4a00a41ed08"). InnerVolumeSpecName "container-storage-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 00:35:40 crc kubenswrapper[4953]: I1203 00:35:40.268054 4953 reconciler_common.go:293] "Volume detached for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/d9fe704d-0942-4735-ac13-d4a00a41ed08-container-storage-run\") on node \"crc\" DevicePath \"\"" Dec 03 00:35:40 crc kubenswrapper[4953]: I1203 00:35:40.372951 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/service-telemetry-framework-index-1-build" event={"ID":"d9fe704d-0942-4735-ac13-d4a00a41ed08","Type":"ContainerDied","Data":"60f494793015bd67b75d60d9080131bc465a708a57502b23672c4c8cdac973d8"} Dec 03 00:35:40 crc kubenswrapper[4953]: I1203 00:35:40.373029 4953 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="60f494793015bd67b75d60d9080131bc465a708a57502b23672c4c8cdac973d8" Dec 03 00:35:40 crc kubenswrapper[4953]: I1203 00:35:40.373150 4953 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/service-telemetry-framework-index-1-build" Dec 03 00:35:41 crc kubenswrapper[4953]: I1203 00:35:41.003040 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d9fe704d-0942-4735-ac13-d4a00a41ed08-container-storage-root" (OuterVolumeSpecName: "container-storage-root") pod "d9fe704d-0942-4735-ac13-d4a00a41ed08" (UID: "d9fe704d-0942-4735-ac13-d4a00a41ed08"). InnerVolumeSpecName "container-storage-root". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 00:35:41 crc kubenswrapper[4953]: I1203 00:35:41.094526 4953 reconciler_common.go:293] "Volume detached for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/d9fe704d-0942-4735-ac13-d4a00a41ed08-container-storage-root\") on node \"crc\" DevicePath \"\"" Dec 03 00:35:42 crc kubenswrapper[4953]: I1203 00:35:42.730294 4953 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["service-telemetry/infrawatch-operators-k89h8"] Dec 03 00:35:42 crc kubenswrapper[4953]: E1203 00:35:42.731198 4953 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d9fe704d-0942-4735-ac13-d4a00a41ed08" containerName="manage-dockerfile" Dec 03 00:35:42 crc kubenswrapper[4953]: I1203 00:35:42.731215 4953 state_mem.go:107] "Deleted CPUSet assignment" podUID="d9fe704d-0942-4735-ac13-d4a00a41ed08" containerName="manage-dockerfile" Dec 03 00:35:42 crc kubenswrapper[4953]: E1203 00:35:42.731241 4953 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d9fe704d-0942-4735-ac13-d4a00a41ed08" containerName="docker-build" Dec 03 00:35:42 crc kubenswrapper[4953]: I1203 00:35:42.731247 4953 state_mem.go:107] "Deleted CPUSet assignment" podUID="d9fe704d-0942-4735-ac13-d4a00a41ed08" containerName="docker-build" Dec 03 00:35:42 crc kubenswrapper[4953]: E1203 00:35:42.731263 4953 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d9fe704d-0942-4735-ac13-d4a00a41ed08" containerName="git-clone" Dec 03 00:35:42 crc kubenswrapper[4953]: I1203 00:35:42.731271 4953 state_mem.go:107] "Deleted CPUSet assignment" podUID="d9fe704d-0942-4735-ac13-d4a00a41ed08" containerName="git-clone" Dec 03 00:35:42 crc kubenswrapper[4953]: I1203 00:35:42.731399 4953 memory_manager.go:354] "RemoveStaleState removing state" podUID="d9fe704d-0942-4735-ac13-d4a00a41ed08" containerName="docker-build" Dec 03 00:35:42 crc kubenswrapper[4953]: I1203 00:35:42.732017 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/infrawatch-operators-k89h8" Dec 03 00:35:42 crc kubenswrapper[4953]: I1203 00:35:42.735274 4953 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"infrawatch-operators-dockercfg-pcc75" Dec 03 00:35:42 crc kubenswrapper[4953]: I1203 00:35:42.751670 4953 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/infrawatch-operators-k89h8"] Dec 03 00:35:42 crc kubenswrapper[4953]: I1203 00:35:42.825908 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g777h\" (UniqueName: \"kubernetes.io/projected/0d237b83-0d72-4651-986c-d97374cab331-kube-api-access-g777h\") pod \"infrawatch-operators-k89h8\" (UID: \"0d237b83-0d72-4651-986c-d97374cab331\") " pod="service-telemetry/infrawatch-operators-k89h8" Dec 03 00:35:42 crc kubenswrapper[4953]: I1203 00:35:42.927860 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g777h\" (UniqueName: \"kubernetes.io/projected/0d237b83-0d72-4651-986c-d97374cab331-kube-api-access-g777h\") pod \"infrawatch-operators-k89h8\" (UID: \"0d237b83-0d72-4651-986c-d97374cab331\") " pod="service-telemetry/infrawatch-operators-k89h8" Dec 03 00:35:42 crc kubenswrapper[4953]: I1203 00:35:42.950282 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g777h\" (UniqueName: \"kubernetes.io/projected/0d237b83-0d72-4651-986c-d97374cab331-kube-api-access-g777h\") pod \"infrawatch-operators-k89h8\" (UID: \"0d237b83-0d72-4651-986c-d97374cab331\") " pod="service-telemetry/infrawatch-operators-k89h8" Dec 03 00:35:43 crc kubenswrapper[4953]: I1203 00:35:43.048983 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/infrawatch-operators-k89h8" Dec 03 00:35:43 crc kubenswrapper[4953]: I1203 00:35:43.530552 4953 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/infrawatch-operators-k89h8"] Dec 03 00:35:44 crc kubenswrapper[4953]: I1203 00:35:44.406918 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/infrawatch-operators-k89h8" event={"ID":"0d237b83-0d72-4651-986c-d97374cab331","Type":"ContainerStarted","Data":"56a80ce8797eccf0799cfb1298f2cd227b1a856fc3dfaa5aeb705c176d29904f"} Dec 03 00:35:54 crc kubenswrapper[4953]: I1203 00:35:54.348469 4953 scope.go:117] "RemoveContainer" containerID="150ffc325f215c6da77fc72304c13cb10dafc800bbc932afd319cc34442a3f76" Dec 03 00:35:54 crc kubenswrapper[4953]: E1203 00:35:54.349841 4953 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6dfbn_openshift-machine-config-operator(53ffb1e9-9a2e-4869-80d7-776f4aa7ef8b)\"" pod="openshift-machine-config-operator/machine-config-daemon-6dfbn" podUID="53ffb1e9-9a2e-4869-80d7-776f4aa7ef8b" Dec 03 00:35:56 crc kubenswrapper[4953]: I1203 00:35:56.579359 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/infrawatch-operators-k89h8" event={"ID":"0d237b83-0d72-4651-986c-d97374cab331","Type":"ContainerStarted","Data":"1344b8b5202a47771c0a40857de9d05ccab7a7e30962e8d02dfd474df9e692a6"} Dec 03 00:35:56 crc kubenswrapper[4953]: I1203 00:35:56.598057 4953 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="service-telemetry/infrawatch-operators-k89h8" podStartSLOduration=2.156763094 podStartE2EDuration="14.598016984s" podCreationTimestamp="2025-12-03 00:35:42 +0000 UTC" firstStartedPulling="2025-12-03 00:35:43.541614219 +0000 UTC m=+1732.525895002" lastFinishedPulling="2025-12-03 00:35:55.982868109 +0000 UTC m=+1744.967148892" observedRunningTime="2025-12-03 00:35:56.596066756 +0000 UTC m=+1745.580347559" watchObservedRunningTime="2025-12-03 00:35:56.598016984 +0000 UTC m=+1745.582297767" Dec 03 00:36:03 crc kubenswrapper[4953]: I1203 00:36:03.049634 4953 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="service-telemetry/infrawatch-operators-k89h8" Dec 03 00:36:03 crc kubenswrapper[4953]: I1203 00:36:03.052096 4953 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="service-telemetry/infrawatch-operators-k89h8" Dec 03 00:36:03 crc kubenswrapper[4953]: I1203 00:36:03.077810 4953 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="service-telemetry/infrawatch-operators-k89h8" Dec 03 00:36:03 crc kubenswrapper[4953]: I1203 00:36:03.661665 4953 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="service-telemetry/infrawatch-operators-k89h8" Dec 03 00:36:07 crc kubenswrapper[4953]: I1203 00:36:07.347929 4953 scope.go:117] "RemoveContainer" containerID="150ffc325f215c6da77fc72304c13cb10dafc800bbc932afd319cc34442a3f76" Dec 03 00:36:07 crc kubenswrapper[4953]: E1203 00:36:07.348768 4953 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6dfbn_openshift-machine-config-operator(53ffb1e9-9a2e-4869-80d7-776f4aa7ef8b)\"" pod="openshift-machine-config-operator/machine-config-daemon-6dfbn" podUID="53ffb1e9-9a2e-4869-80d7-776f4aa7ef8b" Dec 03 00:36:15 crc kubenswrapper[4953]: I1203 00:36:15.577709 4953 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["service-telemetry/372e7d5daac88c2e9a91443a2f508c8c20ad57bc41b1606ec960d61c099m8mf"] Dec 03 00:36:15 crc kubenswrapper[4953]: I1203 00:36:15.580814 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/372e7d5daac88c2e9a91443a2f508c8c20ad57bc41b1606ec960d61c099m8mf" Dec 03 00:36:15 crc kubenswrapper[4953]: I1203 00:36:15.590023 4953 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/372e7d5daac88c2e9a91443a2f508c8c20ad57bc41b1606ec960d61c099m8mf"] Dec 03 00:36:15 crc kubenswrapper[4953]: I1203 00:36:15.752413 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/491c12d8-f039-4fd5-ad84-84ad38e4169f-util\") pod \"372e7d5daac88c2e9a91443a2f508c8c20ad57bc41b1606ec960d61c099m8mf\" (UID: \"491c12d8-f039-4fd5-ad84-84ad38e4169f\") " pod="service-telemetry/372e7d5daac88c2e9a91443a2f508c8c20ad57bc41b1606ec960d61c099m8mf" Dec 03 00:36:15 crc kubenswrapper[4953]: I1203 00:36:15.752843 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/491c12d8-f039-4fd5-ad84-84ad38e4169f-bundle\") pod \"372e7d5daac88c2e9a91443a2f508c8c20ad57bc41b1606ec960d61c099m8mf\" (UID: \"491c12d8-f039-4fd5-ad84-84ad38e4169f\") " pod="service-telemetry/372e7d5daac88c2e9a91443a2f508c8c20ad57bc41b1606ec960d61c099m8mf" Dec 03 00:36:15 crc kubenswrapper[4953]: I1203 00:36:15.753043 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cnxpk\" (UniqueName: \"kubernetes.io/projected/491c12d8-f039-4fd5-ad84-84ad38e4169f-kube-api-access-cnxpk\") pod \"372e7d5daac88c2e9a91443a2f508c8c20ad57bc41b1606ec960d61c099m8mf\" (UID: \"491c12d8-f039-4fd5-ad84-84ad38e4169f\") " pod="service-telemetry/372e7d5daac88c2e9a91443a2f508c8c20ad57bc41b1606ec960d61c099m8mf" Dec 03 00:36:15 crc kubenswrapper[4953]: I1203 00:36:15.854926 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/491c12d8-f039-4fd5-ad84-84ad38e4169f-util\") pod \"372e7d5daac88c2e9a91443a2f508c8c20ad57bc41b1606ec960d61c099m8mf\" (UID: \"491c12d8-f039-4fd5-ad84-84ad38e4169f\") " pod="service-telemetry/372e7d5daac88c2e9a91443a2f508c8c20ad57bc41b1606ec960d61c099m8mf" Dec 03 00:36:15 crc kubenswrapper[4953]: I1203 00:36:15.855045 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/491c12d8-f039-4fd5-ad84-84ad38e4169f-bundle\") pod \"372e7d5daac88c2e9a91443a2f508c8c20ad57bc41b1606ec960d61c099m8mf\" (UID: \"491c12d8-f039-4fd5-ad84-84ad38e4169f\") " pod="service-telemetry/372e7d5daac88c2e9a91443a2f508c8c20ad57bc41b1606ec960d61c099m8mf" Dec 03 00:36:15 crc kubenswrapper[4953]: I1203 00:36:15.855175 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cnxpk\" (UniqueName: \"kubernetes.io/projected/491c12d8-f039-4fd5-ad84-84ad38e4169f-kube-api-access-cnxpk\") pod \"372e7d5daac88c2e9a91443a2f508c8c20ad57bc41b1606ec960d61c099m8mf\" (UID: \"491c12d8-f039-4fd5-ad84-84ad38e4169f\") " pod="service-telemetry/372e7d5daac88c2e9a91443a2f508c8c20ad57bc41b1606ec960d61c099m8mf" Dec 03 00:36:15 crc kubenswrapper[4953]: I1203 00:36:15.855675 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/491c12d8-f039-4fd5-ad84-84ad38e4169f-util\") pod \"372e7d5daac88c2e9a91443a2f508c8c20ad57bc41b1606ec960d61c099m8mf\" (UID: \"491c12d8-f039-4fd5-ad84-84ad38e4169f\") " pod="service-telemetry/372e7d5daac88c2e9a91443a2f508c8c20ad57bc41b1606ec960d61c099m8mf" Dec 03 00:36:15 crc kubenswrapper[4953]: I1203 00:36:15.855774 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/491c12d8-f039-4fd5-ad84-84ad38e4169f-bundle\") pod \"372e7d5daac88c2e9a91443a2f508c8c20ad57bc41b1606ec960d61c099m8mf\" (UID: \"491c12d8-f039-4fd5-ad84-84ad38e4169f\") " pod="service-telemetry/372e7d5daac88c2e9a91443a2f508c8c20ad57bc41b1606ec960d61c099m8mf" Dec 03 00:36:15 crc kubenswrapper[4953]: I1203 00:36:15.878583 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cnxpk\" (UniqueName: \"kubernetes.io/projected/491c12d8-f039-4fd5-ad84-84ad38e4169f-kube-api-access-cnxpk\") pod \"372e7d5daac88c2e9a91443a2f508c8c20ad57bc41b1606ec960d61c099m8mf\" (UID: \"491c12d8-f039-4fd5-ad84-84ad38e4169f\") " pod="service-telemetry/372e7d5daac88c2e9a91443a2f508c8c20ad57bc41b1606ec960d61c099m8mf" Dec 03 00:36:15 crc kubenswrapper[4953]: I1203 00:36:15.909298 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/372e7d5daac88c2e9a91443a2f508c8c20ad57bc41b1606ec960d61c099m8mf" Dec 03 00:36:16 crc kubenswrapper[4953]: I1203 00:36:16.144452 4953 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/372e7d5daac88c2e9a91443a2f508c8c20ad57bc41b1606ec960d61c099m8mf"] Dec 03 00:36:16 crc kubenswrapper[4953]: I1203 00:36:16.378570 4953 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["service-telemetry/500c4f010310dad14c569d8fa2124fef1cf701af50ed1128cec4daf65ab55d5"] Dec 03 00:36:16 crc kubenswrapper[4953]: I1203 00:36:16.390237 4953 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/500c4f010310dad14c569d8fa2124fef1cf701af50ed1128cec4daf65ab55d5"] Dec 03 00:36:16 crc kubenswrapper[4953]: I1203 00:36:16.390522 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/500c4f010310dad14c569d8fa2124fef1cf701af50ed1128cec4daf65ab55d5" Dec 03 00:36:16 crc kubenswrapper[4953]: E1203 00:36:16.499940 4953 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod491c12d8_f039_4fd5_ad84_84ad38e4169f.slice/crio-conmon-9073a3ae806360a54375caf481508460572994ef803c73576bfbf6e78694b6b5.scope\": RecentStats: unable to find data in memory cache]" Dec 03 00:36:16 crc kubenswrapper[4953]: I1203 00:36:16.575819 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/7b9784cd-79d7-45ca-8919-72ae505124ef-util\") pod \"500c4f010310dad14c569d8fa2124fef1cf701af50ed1128cec4daf65ab55d5\" (UID: \"7b9784cd-79d7-45ca-8919-72ae505124ef\") " pod="service-telemetry/500c4f010310dad14c569d8fa2124fef1cf701af50ed1128cec4daf65ab55d5" Dec 03 00:36:16 crc kubenswrapper[4953]: I1203 00:36:16.575893 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/7b9784cd-79d7-45ca-8919-72ae505124ef-bundle\") pod \"500c4f010310dad14c569d8fa2124fef1cf701af50ed1128cec4daf65ab55d5\" (UID: \"7b9784cd-79d7-45ca-8919-72ae505124ef\") " pod="service-telemetry/500c4f010310dad14c569d8fa2124fef1cf701af50ed1128cec4daf65ab55d5" Dec 03 00:36:16 crc kubenswrapper[4953]: I1203 00:36:16.575950 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gzrrc\" (UniqueName: \"kubernetes.io/projected/7b9784cd-79d7-45ca-8919-72ae505124ef-kube-api-access-gzrrc\") pod \"500c4f010310dad14c569d8fa2124fef1cf701af50ed1128cec4daf65ab55d5\" (UID: \"7b9784cd-79d7-45ca-8919-72ae505124ef\") " pod="service-telemetry/500c4f010310dad14c569d8fa2124fef1cf701af50ed1128cec4daf65ab55d5" Dec 03 00:36:16 crc kubenswrapper[4953]: I1203 00:36:16.677249 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/7b9784cd-79d7-45ca-8919-72ae505124ef-util\") pod \"500c4f010310dad14c569d8fa2124fef1cf701af50ed1128cec4daf65ab55d5\" (UID: \"7b9784cd-79d7-45ca-8919-72ae505124ef\") " pod="service-telemetry/500c4f010310dad14c569d8fa2124fef1cf701af50ed1128cec4daf65ab55d5" Dec 03 00:36:16 crc kubenswrapper[4953]: I1203 00:36:16.679050 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/7b9784cd-79d7-45ca-8919-72ae505124ef-bundle\") pod \"500c4f010310dad14c569d8fa2124fef1cf701af50ed1128cec4daf65ab55d5\" (UID: \"7b9784cd-79d7-45ca-8919-72ae505124ef\") " pod="service-telemetry/500c4f010310dad14c569d8fa2124fef1cf701af50ed1128cec4daf65ab55d5" Dec 03 00:36:16 crc kubenswrapper[4953]: I1203 00:36:16.678187 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/7b9784cd-79d7-45ca-8919-72ae505124ef-util\") pod \"500c4f010310dad14c569d8fa2124fef1cf701af50ed1128cec4daf65ab55d5\" (UID: \"7b9784cd-79d7-45ca-8919-72ae505124ef\") " pod="service-telemetry/500c4f010310dad14c569d8fa2124fef1cf701af50ed1128cec4daf65ab55d5" Dec 03 00:36:16 crc kubenswrapper[4953]: I1203 00:36:16.679334 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gzrrc\" (UniqueName: \"kubernetes.io/projected/7b9784cd-79d7-45ca-8919-72ae505124ef-kube-api-access-gzrrc\") pod \"500c4f010310dad14c569d8fa2124fef1cf701af50ed1128cec4daf65ab55d5\" (UID: \"7b9784cd-79d7-45ca-8919-72ae505124ef\") " pod="service-telemetry/500c4f010310dad14c569d8fa2124fef1cf701af50ed1128cec4daf65ab55d5" Dec 03 00:36:16 crc kubenswrapper[4953]: I1203 00:36:16.680077 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/7b9784cd-79d7-45ca-8919-72ae505124ef-bundle\") pod \"500c4f010310dad14c569d8fa2124fef1cf701af50ed1128cec4daf65ab55d5\" (UID: \"7b9784cd-79d7-45ca-8919-72ae505124ef\") " pod="service-telemetry/500c4f010310dad14c569d8fa2124fef1cf701af50ed1128cec4daf65ab55d5" Dec 03 00:36:16 crc kubenswrapper[4953]: I1203 00:36:16.708313 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gzrrc\" (UniqueName: \"kubernetes.io/projected/7b9784cd-79d7-45ca-8919-72ae505124ef-kube-api-access-gzrrc\") pod \"500c4f010310dad14c569d8fa2124fef1cf701af50ed1128cec4daf65ab55d5\" (UID: \"7b9784cd-79d7-45ca-8919-72ae505124ef\") " pod="service-telemetry/500c4f010310dad14c569d8fa2124fef1cf701af50ed1128cec4daf65ab55d5" Dec 03 00:36:16 crc kubenswrapper[4953]: I1203 00:36:16.712072 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/500c4f010310dad14c569d8fa2124fef1cf701af50ed1128cec4daf65ab55d5" Dec 03 00:36:16 crc kubenswrapper[4953]: I1203 00:36:16.749820 4953 generic.go:334] "Generic (PLEG): container finished" podID="491c12d8-f039-4fd5-ad84-84ad38e4169f" containerID="9073a3ae806360a54375caf481508460572994ef803c73576bfbf6e78694b6b5" exitCode=0 Dec 03 00:36:16 crc kubenswrapper[4953]: I1203 00:36:16.749892 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/372e7d5daac88c2e9a91443a2f508c8c20ad57bc41b1606ec960d61c099m8mf" event={"ID":"491c12d8-f039-4fd5-ad84-84ad38e4169f","Type":"ContainerDied","Data":"9073a3ae806360a54375caf481508460572994ef803c73576bfbf6e78694b6b5"} Dec 03 00:36:16 crc kubenswrapper[4953]: I1203 00:36:16.749934 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/372e7d5daac88c2e9a91443a2f508c8c20ad57bc41b1606ec960d61c099m8mf" event={"ID":"491c12d8-f039-4fd5-ad84-84ad38e4169f","Type":"ContainerStarted","Data":"9aea16c7dfb64fa57958f4d875401502165baaafd7a00079c4d1eb0d30e0e43e"} Dec 03 00:36:17 crc kubenswrapper[4953]: I1203 00:36:17.002221 4953 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/500c4f010310dad14c569d8fa2124fef1cf701af50ed1128cec4daf65ab55d5"] Dec 03 00:36:17 crc kubenswrapper[4953]: I1203 00:36:17.761671 4953 generic.go:334] "Generic (PLEG): container finished" podID="491c12d8-f039-4fd5-ad84-84ad38e4169f" containerID="8ba2fa958f0ca09cd3d7446a0f140343df1e2f1d5b1dd18d1cdb51010d5f2a51" exitCode=0 Dec 03 00:36:17 crc kubenswrapper[4953]: I1203 00:36:17.761745 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/372e7d5daac88c2e9a91443a2f508c8c20ad57bc41b1606ec960d61c099m8mf" event={"ID":"491c12d8-f039-4fd5-ad84-84ad38e4169f","Type":"ContainerDied","Data":"8ba2fa958f0ca09cd3d7446a0f140343df1e2f1d5b1dd18d1cdb51010d5f2a51"} Dec 03 00:36:17 crc kubenswrapper[4953]: I1203 00:36:17.770927 4953 generic.go:334] "Generic (PLEG): container finished" podID="7b9784cd-79d7-45ca-8919-72ae505124ef" containerID="be181a020075b75dd5f97a641c0efb91660ac011d57ac79d06be5d63e735f453" exitCode=0 Dec 03 00:36:17 crc kubenswrapper[4953]: I1203 00:36:17.771030 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/500c4f010310dad14c569d8fa2124fef1cf701af50ed1128cec4daf65ab55d5" event={"ID":"7b9784cd-79d7-45ca-8919-72ae505124ef","Type":"ContainerDied","Data":"be181a020075b75dd5f97a641c0efb91660ac011d57ac79d06be5d63e735f453"} Dec 03 00:36:17 crc kubenswrapper[4953]: I1203 00:36:17.771248 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/500c4f010310dad14c569d8fa2124fef1cf701af50ed1128cec4daf65ab55d5" event={"ID":"7b9784cd-79d7-45ca-8919-72ae505124ef","Type":"ContainerStarted","Data":"7ae53997ec449a6892489c5411fbe28e9f39faa80dbbd2d95c4fbcc2cc087834"} Dec 03 00:36:18 crc kubenswrapper[4953]: I1203 00:36:18.782421 4953 generic.go:334] "Generic (PLEG): container finished" podID="491c12d8-f039-4fd5-ad84-84ad38e4169f" containerID="da0a747b00d67cc513617e66109fe967a4ff4ae0e00dac9b19bbf098864b16a5" exitCode=0 Dec 03 00:36:18 crc kubenswrapper[4953]: I1203 00:36:18.783329 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/372e7d5daac88c2e9a91443a2f508c8c20ad57bc41b1606ec960d61c099m8mf" event={"ID":"491c12d8-f039-4fd5-ad84-84ad38e4169f","Type":"ContainerDied","Data":"da0a747b00d67cc513617e66109fe967a4ff4ae0e00dac9b19bbf098864b16a5"} Dec 03 00:36:18 crc kubenswrapper[4953]: I1203 00:36:18.788487 4953 generic.go:334] "Generic (PLEG): container finished" podID="7b9784cd-79d7-45ca-8919-72ae505124ef" containerID="8960d564261213139e50051d4304c909a255952e01da02397d24985f02e0f5b8" exitCode=0 Dec 03 00:36:18 crc kubenswrapper[4953]: I1203 00:36:18.788628 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/500c4f010310dad14c569d8fa2124fef1cf701af50ed1128cec4daf65ab55d5" event={"ID":"7b9784cd-79d7-45ca-8919-72ae505124ef","Type":"ContainerDied","Data":"8960d564261213139e50051d4304c909a255952e01da02397d24985f02e0f5b8"} Dec 03 00:36:19 crc kubenswrapper[4953]: I1203 00:36:19.348355 4953 scope.go:117] "RemoveContainer" containerID="150ffc325f215c6da77fc72304c13cb10dafc800bbc932afd319cc34442a3f76" Dec 03 00:36:19 crc kubenswrapper[4953]: E1203 00:36:19.349241 4953 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6dfbn_openshift-machine-config-operator(53ffb1e9-9a2e-4869-80d7-776f4aa7ef8b)\"" pod="openshift-machine-config-operator/machine-config-daemon-6dfbn" podUID="53ffb1e9-9a2e-4869-80d7-776f4aa7ef8b" Dec 03 00:36:19 crc kubenswrapper[4953]: I1203 00:36:19.801377 4953 generic.go:334] "Generic (PLEG): container finished" podID="7b9784cd-79d7-45ca-8919-72ae505124ef" containerID="c7de610f87d1ef61ffc7a2e80b3d84884ff4ab55213a39f3a29865ec0279cb61" exitCode=0 Dec 03 00:36:19 crc kubenswrapper[4953]: I1203 00:36:19.801489 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/500c4f010310dad14c569d8fa2124fef1cf701af50ed1128cec4daf65ab55d5" event={"ID":"7b9784cd-79d7-45ca-8919-72ae505124ef","Type":"ContainerDied","Data":"c7de610f87d1ef61ffc7a2e80b3d84884ff4ab55213a39f3a29865ec0279cb61"} Dec 03 00:36:20 crc kubenswrapper[4953]: I1203 00:36:20.064842 4953 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/372e7d5daac88c2e9a91443a2f508c8c20ad57bc41b1606ec960d61c099m8mf" Dec 03 00:36:20 crc kubenswrapper[4953]: I1203 00:36:20.247319 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/491c12d8-f039-4fd5-ad84-84ad38e4169f-util\") pod \"491c12d8-f039-4fd5-ad84-84ad38e4169f\" (UID: \"491c12d8-f039-4fd5-ad84-84ad38e4169f\") " Dec 03 00:36:20 crc kubenswrapper[4953]: I1203 00:36:20.247530 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cnxpk\" (UniqueName: \"kubernetes.io/projected/491c12d8-f039-4fd5-ad84-84ad38e4169f-kube-api-access-cnxpk\") pod \"491c12d8-f039-4fd5-ad84-84ad38e4169f\" (UID: \"491c12d8-f039-4fd5-ad84-84ad38e4169f\") " Dec 03 00:36:20 crc kubenswrapper[4953]: I1203 00:36:20.247583 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/491c12d8-f039-4fd5-ad84-84ad38e4169f-bundle\") pod \"491c12d8-f039-4fd5-ad84-84ad38e4169f\" (UID: \"491c12d8-f039-4fd5-ad84-84ad38e4169f\") " Dec 03 00:36:20 crc kubenswrapper[4953]: I1203 00:36:20.248522 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/491c12d8-f039-4fd5-ad84-84ad38e4169f-bundle" (OuterVolumeSpecName: "bundle") pod "491c12d8-f039-4fd5-ad84-84ad38e4169f" (UID: "491c12d8-f039-4fd5-ad84-84ad38e4169f"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 00:36:20 crc kubenswrapper[4953]: I1203 00:36:20.255324 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/491c12d8-f039-4fd5-ad84-84ad38e4169f-kube-api-access-cnxpk" (OuterVolumeSpecName: "kube-api-access-cnxpk") pod "491c12d8-f039-4fd5-ad84-84ad38e4169f" (UID: "491c12d8-f039-4fd5-ad84-84ad38e4169f"). InnerVolumeSpecName "kube-api-access-cnxpk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:36:20 crc kubenswrapper[4953]: I1203 00:36:20.270218 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/491c12d8-f039-4fd5-ad84-84ad38e4169f-util" (OuterVolumeSpecName: "util") pod "491c12d8-f039-4fd5-ad84-84ad38e4169f" (UID: "491c12d8-f039-4fd5-ad84-84ad38e4169f"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 00:36:20 crc kubenswrapper[4953]: I1203 00:36:20.349055 4953 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/491c12d8-f039-4fd5-ad84-84ad38e4169f-util\") on node \"crc\" DevicePath \"\"" Dec 03 00:36:20 crc kubenswrapper[4953]: I1203 00:36:20.349115 4953 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cnxpk\" (UniqueName: \"kubernetes.io/projected/491c12d8-f039-4fd5-ad84-84ad38e4169f-kube-api-access-cnxpk\") on node \"crc\" DevicePath \"\"" Dec 03 00:36:20 crc kubenswrapper[4953]: I1203 00:36:20.349130 4953 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/491c12d8-f039-4fd5-ad84-84ad38e4169f-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 00:36:20 crc kubenswrapper[4953]: I1203 00:36:20.811737 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/372e7d5daac88c2e9a91443a2f508c8c20ad57bc41b1606ec960d61c099m8mf" event={"ID":"491c12d8-f039-4fd5-ad84-84ad38e4169f","Type":"ContainerDied","Data":"9aea16c7dfb64fa57958f4d875401502165baaafd7a00079c4d1eb0d30e0e43e"} Dec 03 00:36:20 crc kubenswrapper[4953]: I1203 00:36:20.811805 4953 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9aea16c7dfb64fa57958f4d875401502165baaafd7a00079c4d1eb0d30e0e43e" Dec 03 00:36:20 crc kubenswrapper[4953]: I1203 00:36:20.811830 4953 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/372e7d5daac88c2e9a91443a2f508c8c20ad57bc41b1606ec960d61c099m8mf" Dec 03 00:36:21 crc kubenswrapper[4953]: I1203 00:36:21.072360 4953 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/500c4f010310dad14c569d8fa2124fef1cf701af50ed1128cec4daf65ab55d5" Dec 03 00:36:21 crc kubenswrapper[4953]: I1203 00:36:21.264458 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/7b9784cd-79d7-45ca-8919-72ae505124ef-util\") pod \"7b9784cd-79d7-45ca-8919-72ae505124ef\" (UID: \"7b9784cd-79d7-45ca-8919-72ae505124ef\") " Dec 03 00:36:21 crc kubenswrapper[4953]: I1203 00:36:21.264868 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gzrrc\" (UniqueName: \"kubernetes.io/projected/7b9784cd-79d7-45ca-8919-72ae505124ef-kube-api-access-gzrrc\") pod \"7b9784cd-79d7-45ca-8919-72ae505124ef\" (UID: \"7b9784cd-79d7-45ca-8919-72ae505124ef\") " Dec 03 00:36:21 crc kubenswrapper[4953]: I1203 00:36:21.264894 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/7b9784cd-79d7-45ca-8919-72ae505124ef-bundle\") pod \"7b9784cd-79d7-45ca-8919-72ae505124ef\" (UID: \"7b9784cd-79d7-45ca-8919-72ae505124ef\") " Dec 03 00:36:21 crc kubenswrapper[4953]: I1203 00:36:21.267388 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7b9784cd-79d7-45ca-8919-72ae505124ef-bundle" (OuterVolumeSpecName: "bundle") pod "7b9784cd-79d7-45ca-8919-72ae505124ef" (UID: "7b9784cd-79d7-45ca-8919-72ae505124ef"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 00:36:21 crc kubenswrapper[4953]: I1203 00:36:21.272177 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7b9784cd-79d7-45ca-8919-72ae505124ef-kube-api-access-gzrrc" (OuterVolumeSpecName: "kube-api-access-gzrrc") pod "7b9784cd-79d7-45ca-8919-72ae505124ef" (UID: "7b9784cd-79d7-45ca-8919-72ae505124ef"). InnerVolumeSpecName "kube-api-access-gzrrc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:36:21 crc kubenswrapper[4953]: I1203 00:36:21.286710 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7b9784cd-79d7-45ca-8919-72ae505124ef-util" (OuterVolumeSpecName: "util") pod "7b9784cd-79d7-45ca-8919-72ae505124ef" (UID: "7b9784cd-79d7-45ca-8919-72ae505124ef"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 00:36:21 crc kubenswrapper[4953]: I1203 00:36:21.366910 4953 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/7b9784cd-79d7-45ca-8919-72ae505124ef-util\") on node \"crc\" DevicePath \"\"" Dec 03 00:36:21 crc kubenswrapper[4953]: I1203 00:36:21.366947 4953 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gzrrc\" (UniqueName: \"kubernetes.io/projected/7b9784cd-79d7-45ca-8919-72ae505124ef-kube-api-access-gzrrc\") on node \"crc\" DevicePath \"\"" Dec 03 00:36:21 crc kubenswrapper[4953]: I1203 00:36:21.366963 4953 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/7b9784cd-79d7-45ca-8919-72ae505124ef-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 00:36:21 crc kubenswrapper[4953]: I1203 00:36:21.822638 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/500c4f010310dad14c569d8fa2124fef1cf701af50ed1128cec4daf65ab55d5" event={"ID":"7b9784cd-79d7-45ca-8919-72ae505124ef","Type":"ContainerDied","Data":"7ae53997ec449a6892489c5411fbe28e9f39faa80dbbd2d95c4fbcc2cc087834"} Dec 03 00:36:21 crc kubenswrapper[4953]: I1203 00:36:21.822680 4953 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7ae53997ec449a6892489c5411fbe28e9f39faa80dbbd2d95c4fbcc2cc087834" Dec 03 00:36:21 crc kubenswrapper[4953]: I1203 00:36:21.822757 4953 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/500c4f010310dad14c569d8fa2124fef1cf701af50ed1128cec4daf65ab55d5" Dec 03 00:36:25 crc kubenswrapper[4953]: I1203 00:36:25.486370 4953 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["service-telemetry/service-telemetry-operator-75f4bf7844-lfwk2"] Dec 03 00:36:25 crc kubenswrapper[4953]: E1203 00:36:25.487270 4953 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="491c12d8-f039-4fd5-ad84-84ad38e4169f" containerName="extract" Dec 03 00:36:25 crc kubenswrapper[4953]: I1203 00:36:25.487288 4953 state_mem.go:107] "Deleted CPUSet assignment" podUID="491c12d8-f039-4fd5-ad84-84ad38e4169f" containerName="extract" Dec 03 00:36:25 crc kubenswrapper[4953]: E1203 00:36:25.487303 4953 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="491c12d8-f039-4fd5-ad84-84ad38e4169f" containerName="pull" Dec 03 00:36:25 crc kubenswrapper[4953]: I1203 00:36:25.487311 4953 state_mem.go:107] "Deleted CPUSet assignment" podUID="491c12d8-f039-4fd5-ad84-84ad38e4169f" containerName="pull" Dec 03 00:36:25 crc kubenswrapper[4953]: E1203 00:36:25.487326 4953 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="491c12d8-f039-4fd5-ad84-84ad38e4169f" containerName="util" Dec 03 00:36:25 crc kubenswrapper[4953]: I1203 00:36:25.487335 4953 state_mem.go:107] "Deleted CPUSet assignment" podUID="491c12d8-f039-4fd5-ad84-84ad38e4169f" containerName="util" Dec 03 00:36:25 crc kubenswrapper[4953]: E1203 00:36:25.487347 4953 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7b9784cd-79d7-45ca-8919-72ae505124ef" containerName="extract" Dec 03 00:36:25 crc kubenswrapper[4953]: I1203 00:36:25.487354 4953 state_mem.go:107] "Deleted CPUSet assignment" podUID="7b9784cd-79d7-45ca-8919-72ae505124ef" containerName="extract" Dec 03 00:36:25 crc kubenswrapper[4953]: E1203 00:36:25.487365 4953 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7b9784cd-79d7-45ca-8919-72ae505124ef" containerName="util" Dec 03 00:36:25 crc kubenswrapper[4953]: I1203 00:36:25.487372 4953 state_mem.go:107] "Deleted CPUSet assignment" podUID="7b9784cd-79d7-45ca-8919-72ae505124ef" containerName="util" Dec 03 00:36:25 crc kubenswrapper[4953]: E1203 00:36:25.487383 4953 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7b9784cd-79d7-45ca-8919-72ae505124ef" containerName="pull" Dec 03 00:36:25 crc kubenswrapper[4953]: I1203 00:36:25.487390 4953 state_mem.go:107] "Deleted CPUSet assignment" podUID="7b9784cd-79d7-45ca-8919-72ae505124ef" containerName="pull" Dec 03 00:36:25 crc kubenswrapper[4953]: I1203 00:36:25.487508 4953 memory_manager.go:354] "RemoveStaleState removing state" podUID="7b9784cd-79d7-45ca-8919-72ae505124ef" containerName="extract" Dec 03 00:36:25 crc kubenswrapper[4953]: I1203 00:36:25.487541 4953 memory_manager.go:354] "RemoveStaleState removing state" podUID="491c12d8-f039-4fd5-ad84-84ad38e4169f" containerName="extract" Dec 03 00:36:25 crc kubenswrapper[4953]: I1203 00:36:25.488204 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/service-telemetry-operator-75f4bf7844-lfwk2" Dec 03 00:36:25 crc kubenswrapper[4953]: I1203 00:36:25.490878 4953 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"service-telemetry-operator-dockercfg-k4sjm" Dec 03 00:36:25 crc kubenswrapper[4953]: I1203 00:36:25.507716 4953 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/service-telemetry-operator-75f4bf7844-lfwk2"] Dec 03 00:36:25 crc kubenswrapper[4953]: I1203 00:36:25.733527 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"runner\" (UniqueName: \"kubernetes.io/empty-dir/8353057a-2770-4006-a75c-37c59f034192-runner\") pod \"service-telemetry-operator-75f4bf7844-lfwk2\" (UID: \"8353057a-2770-4006-a75c-37c59f034192\") " pod="service-telemetry/service-telemetry-operator-75f4bf7844-lfwk2" Dec 03 00:36:25 crc kubenswrapper[4953]: I1203 00:36:25.733581 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bfm5t\" (UniqueName: \"kubernetes.io/projected/8353057a-2770-4006-a75c-37c59f034192-kube-api-access-bfm5t\") pod \"service-telemetry-operator-75f4bf7844-lfwk2\" (UID: \"8353057a-2770-4006-a75c-37c59f034192\") " pod="service-telemetry/service-telemetry-operator-75f4bf7844-lfwk2" Dec 03 00:36:25 crc kubenswrapper[4953]: I1203 00:36:25.836067 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"runner\" (UniqueName: \"kubernetes.io/empty-dir/8353057a-2770-4006-a75c-37c59f034192-runner\") pod \"service-telemetry-operator-75f4bf7844-lfwk2\" (UID: \"8353057a-2770-4006-a75c-37c59f034192\") " pod="service-telemetry/service-telemetry-operator-75f4bf7844-lfwk2" Dec 03 00:36:25 crc kubenswrapper[4953]: I1203 00:36:25.836118 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bfm5t\" (UniqueName: \"kubernetes.io/projected/8353057a-2770-4006-a75c-37c59f034192-kube-api-access-bfm5t\") pod \"service-telemetry-operator-75f4bf7844-lfwk2\" (UID: \"8353057a-2770-4006-a75c-37c59f034192\") " pod="service-telemetry/service-telemetry-operator-75f4bf7844-lfwk2" Dec 03 00:36:25 crc kubenswrapper[4953]: I1203 00:36:25.836878 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"runner\" (UniqueName: \"kubernetes.io/empty-dir/8353057a-2770-4006-a75c-37c59f034192-runner\") pod \"service-telemetry-operator-75f4bf7844-lfwk2\" (UID: \"8353057a-2770-4006-a75c-37c59f034192\") " pod="service-telemetry/service-telemetry-operator-75f4bf7844-lfwk2" Dec 03 00:36:25 crc kubenswrapper[4953]: I1203 00:36:25.858820 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bfm5t\" (UniqueName: \"kubernetes.io/projected/8353057a-2770-4006-a75c-37c59f034192-kube-api-access-bfm5t\") pod \"service-telemetry-operator-75f4bf7844-lfwk2\" (UID: \"8353057a-2770-4006-a75c-37c59f034192\") " pod="service-telemetry/service-telemetry-operator-75f4bf7844-lfwk2" Dec 03 00:36:26 crc kubenswrapper[4953]: I1203 00:36:26.040991 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/service-telemetry-operator-75f4bf7844-lfwk2" Dec 03 00:36:26 crc kubenswrapper[4953]: I1203 00:36:26.605841 4953 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/service-telemetry-operator-75f4bf7844-lfwk2"] Dec 03 00:36:26 crc kubenswrapper[4953]: W1203 00:36:26.618048 4953 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8353057a_2770_4006_a75c_37c59f034192.slice/crio-6ff3232d88b62e105c0c315437363ed18cb36b8b938af63d47bc1c8df6cb3d55 WatchSource:0}: Error finding container 6ff3232d88b62e105c0c315437363ed18cb36b8b938af63d47bc1c8df6cb3d55: Status 404 returned error can't find the container with id 6ff3232d88b62e105c0c315437363ed18cb36b8b938af63d47bc1c8df6cb3d55 Dec 03 00:36:26 crc kubenswrapper[4953]: I1203 00:36:26.873170 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/service-telemetry-operator-75f4bf7844-lfwk2" event={"ID":"8353057a-2770-4006-a75c-37c59f034192","Type":"ContainerStarted","Data":"6ff3232d88b62e105c0c315437363ed18cb36b8b938af63d47bc1c8df6cb3d55"} Dec 03 00:36:28 crc kubenswrapper[4953]: I1203 00:36:28.955551 4953 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["service-telemetry/smart-gateway-operator-7cddd7d4d-vnx98"] Dec 03 00:36:28 crc kubenswrapper[4953]: I1203 00:36:28.956317 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/smart-gateway-operator-7cddd7d4d-vnx98" Dec 03 00:36:28 crc kubenswrapper[4953]: I1203 00:36:28.966210 4953 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"smart-gateway-operator-dockercfg-d8v5s" Dec 03 00:36:28 crc kubenswrapper[4953]: I1203 00:36:28.979193 4953 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/smart-gateway-operator-7cddd7d4d-vnx98"] Dec 03 00:36:29 crc kubenswrapper[4953]: I1203 00:36:29.081060 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"runner\" (UniqueName: \"kubernetes.io/empty-dir/34c1166e-eb61-4a66-82bd-6498b4d908a8-runner\") pod \"smart-gateway-operator-7cddd7d4d-vnx98\" (UID: \"34c1166e-eb61-4a66-82bd-6498b4d908a8\") " pod="service-telemetry/smart-gateway-operator-7cddd7d4d-vnx98" Dec 03 00:36:29 crc kubenswrapper[4953]: I1203 00:36:29.081121 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qbqj4\" (UniqueName: \"kubernetes.io/projected/34c1166e-eb61-4a66-82bd-6498b4d908a8-kube-api-access-qbqj4\") pod \"smart-gateway-operator-7cddd7d4d-vnx98\" (UID: \"34c1166e-eb61-4a66-82bd-6498b4d908a8\") " pod="service-telemetry/smart-gateway-operator-7cddd7d4d-vnx98" Dec 03 00:36:29 crc kubenswrapper[4953]: I1203 00:36:29.182501 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"runner\" (UniqueName: \"kubernetes.io/empty-dir/34c1166e-eb61-4a66-82bd-6498b4d908a8-runner\") pod \"smart-gateway-operator-7cddd7d4d-vnx98\" (UID: \"34c1166e-eb61-4a66-82bd-6498b4d908a8\") " pod="service-telemetry/smart-gateway-operator-7cddd7d4d-vnx98" Dec 03 00:36:29 crc kubenswrapper[4953]: I1203 00:36:29.182556 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qbqj4\" (UniqueName: \"kubernetes.io/projected/34c1166e-eb61-4a66-82bd-6498b4d908a8-kube-api-access-qbqj4\") pod \"smart-gateway-operator-7cddd7d4d-vnx98\" (UID: \"34c1166e-eb61-4a66-82bd-6498b4d908a8\") " pod="service-telemetry/smart-gateway-operator-7cddd7d4d-vnx98" Dec 03 00:36:29 crc kubenswrapper[4953]: I1203 00:36:29.183531 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"runner\" (UniqueName: \"kubernetes.io/empty-dir/34c1166e-eb61-4a66-82bd-6498b4d908a8-runner\") pod \"smart-gateway-operator-7cddd7d4d-vnx98\" (UID: \"34c1166e-eb61-4a66-82bd-6498b4d908a8\") " pod="service-telemetry/smart-gateway-operator-7cddd7d4d-vnx98" Dec 03 00:36:29 crc kubenswrapper[4953]: I1203 00:36:29.222703 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qbqj4\" (UniqueName: \"kubernetes.io/projected/34c1166e-eb61-4a66-82bd-6498b4d908a8-kube-api-access-qbqj4\") pod \"smart-gateway-operator-7cddd7d4d-vnx98\" (UID: \"34c1166e-eb61-4a66-82bd-6498b4d908a8\") " pod="service-telemetry/smart-gateway-operator-7cddd7d4d-vnx98" Dec 03 00:36:29 crc kubenswrapper[4953]: I1203 00:36:29.283168 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/smart-gateway-operator-7cddd7d4d-vnx98" Dec 03 00:36:29 crc kubenswrapper[4953]: I1203 00:36:29.646199 4953 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/smart-gateway-operator-7cddd7d4d-vnx98"] Dec 03 00:36:29 crc kubenswrapper[4953]: I1203 00:36:29.656580 4953 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 03 00:36:29 crc kubenswrapper[4953]: I1203 00:36:29.900382 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/smart-gateway-operator-7cddd7d4d-vnx98" event={"ID":"34c1166e-eb61-4a66-82bd-6498b4d908a8","Type":"ContainerStarted","Data":"757d09ec16735a0b71e3aee4d0786c92f5b55dcf722fd2f07116bb695dfc7876"} Dec 03 00:36:32 crc kubenswrapper[4953]: I1203 00:36:32.348288 4953 scope.go:117] "RemoveContainer" containerID="150ffc325f215c6da77fc72304c13cb10dafc800bbc932afd319cc34442a3f76" Dec 03 00:36:32 crc kubenswrapper[4953]: E1203 00:36:32.349235 4953 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6dfbn_openshift-machine-config-operator(53ffb1e9-9a2e-4869-80d7-776f4aa7ef8b)\"" pod="openshift-machine-config-operator/machine-config-daemon-6dfbn" podUID="53ffb1e9-9a2e-4869-80d7-776f4aa7ef8b" Dec 03 00:36:46 crc kubenswrapper[4953]: I1203 00:36:46.348211 4953 scope.go:117] "RemoveContainer" containerID="150ffc325f215c6da77fc72304c13cb10dafc800bbc932afd319cc34442a3f76" Dec 03 00:36:46 crc kubenswrapper[4953]: E1203 00:36:46.350266 4953 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6dfbn_openshift-machine-config-operator(53ffb1e9-9a2e-4869-80d7-776f4aa7ef8b)\"" pod="openshift-machine-config-operator/machine-config-daemon-6dfbn" podUID="53ffb1e9-9a2e-4869-80d7-776f4aa7ef8b" Dec 03 00:36:49 crc kubenswrapper[4953]: E1203 00:36:49.437850 4953 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/infrawatch/smart-gateway-operator:latest" Dec 03 00:36:49 crc kubenswrapper[4953]: E1203 00:36:49.438817 4953 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:operator,Image:quay.io/infrawatch/smart-gateway-operator:latest,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:WATCH_NAMESPACE,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.annotations['olm.targetNamespaces'],},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:POD_NAME,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.name,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:OPERATOR_NAME,Value:smart-gateway-operator,ValueFrom:nil,},EnvVar{Name:ANSIBLE_GATHERING,Value:explicit,ValueFrom:nil,},EnvVar{Name:ANSIBLE_VERBOSITY_SMARTGATEWAY_SMARTGATEWAY_INFRA_WATCH,Value:4,ValueFrom:nil,},EnvVar{Name:ANSIBLE_DEBUG_LOGS,Value:true,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CORE_SMARTGATEWAY_IMAGE,Value:image-registry.openshift-image-registry.svc:5000/service-telemetry/sg-core:latest,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_BRIDGE_SMARTGATEWAY_IMAGE,Value:image-registry.openshift-image-registry.svc:5000/service-telemetry/sg-bridge:latest,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OAUTH_PROXY_IMAGE,Value:quay.io/openshift/origin-oauth-proxy:latest,ValueFrom:nil,},EnvVar{Name:OPERATOR_CONDITION_NAME,Value:smart-gateway-operator.v5.0.1764722025,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:runner,ReadOnly:false,MountPath:/tmp/ansible-operator/runner,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-qbqj4,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000670000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod smart-gateway-operator-7cddd7d4d-vnx98_service-telemetry(34c1166e-eb61-4a66-82bd-6498b4d908a8): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 03 00:36:49 crc kubenswrapper[4953]: E1203 00:36:49.440049 4953 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="service-telemetry/smart-gateway-operator-7cddd7d4d-vnx98" podUID="34c1166e-eb61-4a66-82bd-6498b4d908a8" Dec 03 00:36:49 crc kubenswrapper[4953]: E1203 00:36:49.596214 4953 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/infrawatch/service-telemetry-operator:latest" Dec 03 00:36:49 crc kubenswrapper[4953]: E1203 00:36:49.596433 4953 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:operator,Image:quay.io/infrawatch/service-telemetry-operator:latest,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:WATCH_NAMESPACE,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.annotations['olm.targetNamespaces'],},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:POD_NAME,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.name,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:OPERATOR_NAME,Value:service-telemetry-operator,ValueFrom:nil,},EnvVar{Name:ANSIBLE_GATHERING,Value:explicit,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_PROMETHEUS_WEBHOOK_SNMP_IMAGE,Value:image-registry.openshift-image-registry.svc:5000/service-telemetry/prometheus-webhook-snmp:latest,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OAUTH_PROXY_IMAGE,Value:quay.io/openshift/origin-oauth-proxy:latest,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_PROMETHEUS_IMAGE,Value:quay.io/prometheus/prometheus:latest,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_ALERTMANAGER_IMAGE,Value:quay.io/prometheus/alertmanager:latest,ValueFrom:nil,},EnvVar{Name:OPERATOR_CONDITION_NAME,Value:service-telemetry-operator.v1.5.1764722027,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:runner,ReadOnly:false,MountPath:/tmp/ansible-operator/runner,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-bfm5t,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000670000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod service-telemetry-operator-75f4bf7844-lfwk2_service-telemetry(8353057a-2770-4006-a75c-37c59f034192): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 03 00:36:49 crc kubenswrapper[4953]: E1203 00:36:49.597695 4953 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="service-telemetry/service-telemetry-operator-75f4bf7844-lfwk2" podUID="8353057a-2770-4006-a75c-37c59f034192" Dec 03 00:36:50 crc kubenswrapper[4953]: E1203 00:36:50.088913 4953 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/infrawatch/smart-gateway-operator:latest\\\"\"" pod="service-telemetry/smart-gateway-operator-7cddd7d4d-vnx98" podUID="34c1166e-eb61-4a66-82bd-6498b4d908a8" Dec 03 00:36:50 crc kubenswrapper[4953]: E1203 00:36:50.088927 4953 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/infrawatch/service-telemetry-operator:latest\\\"\"" pod="service-telemetry/service-telemetry-operator-75f4bf7844-lfwk2" podUID="8353057a-2770-4006-a75c-37c59f034192" Dec 03 00:36:57 crc kubenswrapper[4953]: I1203 00:36:57.348152 4953 scope.go:117] "RemoveContainer" containerID="150ffc325f215c6da77fc72304c13cb10dafc800bbc932afd319cc34442a3f76" Dec 03 00:36:57 crc kubenswrapper[4953]: E1203 00:36:57.349523 4953 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6dfbn_openshift-machine-config-operator(53ffb1e9-9a2e-4869-80d7-776f4aa7ef8b)\"" pod="openshift-machine-config-operator/machine-config-daemon-6dfbn" podUID="53ffb1e9-9a2e-4869-80d7-776f4aa7ef8b" Dec 03 00:37:06 crc kubenswrapper[4953]: I1203 00:37:06.218414 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/smart-gateway-operator-7cddd7d4d-vnx98" event={"ID":"34c1166e-eb61-4a66-82bd-6498b4d908a8","Type":"ContainerStarted","Data":"07ecef7852a6831f5d9f1e18bd35baf51b9ebd57c87d30e9de0399a18e24f113"} Dec 03 00:37:06 crc kubenswrapper[4953]: I1203 00:37:06.222425 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/service-telemetry-operator-75f4bf7844-lfwk2" event={"ID":"8353057a-2770-4006-a75c-37c59f034192","Type":"ContainerStarted","Data":"6bb8113ec30666a8902ad7d9ea58032fe976a0b5f98ed9a41b21c7817e144f7a"} Dec 03 00:37:06 crc kubenswrapper[4953]: I1203 00:37:06.254243 4953 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="service-telemetry/smart-gateway-operator-7cddd7d4d-vnx98" podStartSLOduration=2.475625182 podStartE2EDuration="38.254208907s" podCreationTimestamp="2025-12-03 00:36:28 +0000 UTC" firstStartedPulling="2025-12-03 00:36:29.656279795 +0000 UTC m=+1778.640560568" lastFinishedPulling="2025-12-03 00:37:05.43486351 +0000 UTC m=+1814.419144293" observedRunningTime="2025-12-03 00:37:06.243287297 +0000 UTC m=+1815.227568110" watchObservedRunningTime="2025-12-03 00:37:06.254208907 +0000 UTC m=+1815.238489720" Dec 03 00:37:06 crc kubenswrapper[4953]: I1203 00:37:06.278814 4953 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="service-telemetry/service-telemetry-operator-75f4bf7844-lfwk2" podStartSLOduration=2.571682857 podStartE2EDuration="41.278785024s" podCreationTimestamp="2025-12-03 00:36:25 +0000 UTC" firstStartedPulling="2025-12-03 00:36:26.619825597 +0000 UTC m=+1775.604106380" lastFinishedPulling="2025-12-03 00:37:05.326927764 +0000 UTC m=+1814.311208547" observedRunningTime="2025-12-03 00:37:06.267517386 +0000 UTC m=+1815.251798199" watchObservedRunningTime="2025-12-03 00:37:06.278785024 +0000 UTC m=+1815.263065847" Dec 03 00:37:12 crc kubenswrapper[4953]: I1203 00:37:12.348058 4953 scope.go:117] "RemoveContainer" containerID="150ffc325f215c6da77fc72304c13cb10dafc800bbc932afd319cc34442a3f76" Dec 03 00:37:12 crc kubenswrapper[4953]: E1203 00:37:12.350136 4953 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6dfbn_openshift-machine-config-operator(53ffb1e9-9a2e-4869-80d7-776f4aa7ef8b)\"" pod="openshift-machine-config-operator/machine-config-daemon-6dfbn" podUID="53ffb1e9-9a2e-4869-80d7-776f4aa7ef8b" Dec 03 00:37:27 crc kubenswrapper[4953]: I1203 00:37:27.349650 4953 scope.go:117] "RemoveContainer" containerID="150ffc325f215c6da77fc72304c13cb10dafc800bbc932afd319cc34442a3f76" Dec 03 00:37:27 crc kubenswrapper[4953]: E1203 00:37:27.350733 4953 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6dfbn_openshift-machine-config-operator(53ffb1e9-9a2e-4869-80d7-776f4aa7ef8b)\"" pod="openshift-machine-config-operator/machine-config-daemon-6dfbn" podUID="53ffb1e9-9a2e-4869-80d7-776f4aa7ef8b" Dec 03 00:37:35 crc kubenswrapper[4953]: I1203 00:37:35.587671 4953 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["service-telemetry/default-interconnect-68864d46cb-mv5rb"] Dec 03 00:37:35 crc kubenswrapper[4953]: I1203 00:37:35.589631 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/default-interconnect-68864d46cb-mv5rb" Dec 03 00:37:35 crc kubenswrapper[4953]: I1203 00:37:35.592784 4953 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"default-interconnect-users" Dec 03 00:37:35 crc kubenswrapper[4953]: I1203 00:37:35.592926 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"default-interconnect-sasl-config" Dec 03 00:37:35 crc kubenswrapper[4953]: I1203 00:37:35.592962 4953 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"default-interconnect-inter-router-credentials" Dec 03 00:37:35 crc kubenswrapper[4953]: I1203 00:37:35.593198 4953 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"default-interconnect-openstack-credentials" Dec 03 00:37:35 crc kubenswrapper[4953]: I1203 00:37:35.593359 4953 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"default-interconnect-dockercfg-d5qbv" Dec 03 00:37:35 crc kubenswrapper[4953]: I1203 00:37:35.593851 4953 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"default-interconnect-openstack-ca" Dec 03 00:37:35 crc kubenswrapper[4953]: I1203 00:37:35.594837 4953 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"default-interconnect-inter-router-ca" Dec 03 00:37:35 crc kubenswrapper[4953]: I1203 00:37:35.614469 4953 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/default-interconnect-68864d46cb-mv5rb"] Dec 03 00:37:35 crc kubenswrapper[4953]: I1203 00:37:35.627863 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-shh7v\" (UniqueName: \"kubernetes.io/projected/7e7fa27e-b88d-4d1c-b493-63b8f811dff5-kube-api-access-shh7v\") pod \"default-interconnect-68864d46cb-mv5rb\" (UID: \"7e7fa27e-b88d-4d1c-b493-63b8f811dff5\") " pod="service-telemetry/default-interconnect-68864d46cb-mv5rb" Dec 03 00:37:35 crc kubenswrapper[4953]: I1203 00:37:35.627958 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-interconnect-inter-router-credentials\" (UniqueName: \"kubernetes.io/secret/7e7fa27e-b88d-4d1c-b493-63b8f811dff5-default-interconnect-inter-router-credentials\") pod \"default-interconnect-68864d46cb-mv5rb\" (UID: \"7e7fa27e-b88d-4d1c-b493-63b8f811dff5\") " pod="service-telemetry/default-interconnect-68864d46cb-mv5rb" Dec 03 00:37:35 crc kubenswrapper[4953]: I1203 00:37:35.628001 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-interconnect-openstack-credentials\" (UniqueName: \"kubernetes.io/secret/7e7fa27e-b88d-4d1c-b493-63b8f811dff5-default-interconnect-openstack-credentials\") pod \"default-interconnect-68864d46cb-mv5rb\" (UID: \"7e7fa27e-b88d-4d1c-b493-63b8f811dff5\") " pod="service-telemetry/default-interconnect-68864d46cb-mv5rb" Dec 03 00:37:35 crc kubenswrapper[4953]: I1203 00:37:35.628024 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-interconnect-inter-router-ca\" (UniqueName: \"kubernetes.io/secret/7e7fa27e-b88d-4d1c-b493-63b8f811dff5-default-interconnect-inter-router-ca\") pod \"default-interconnect-68864d46cb-mv5rb\" (UID: \"7e7fa27e-b88d-4d1c-b493-63b8f811dff5\") " pod="service-telemetry/default-interconnect-68864d46cb-mv5rb" Dec 03 00:37:35 crc kubenswrapper[4953]: I1203 00:37:35.628057 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sasl-config\" (UniqueName: \"kubernetes.io/configmap/7e7fa27e-b88d-4d1c-b493-63b8f811dff5-sasl-config\") pod \"default-interconnect-68864d46cb-mv5rb\" (UID: \"7e7fa27e-b88d-4d1c-b493-63b8f811dff5\") " pod="service-telemetry/default-interconnect-68864d46cb-mv5rb" Dec 03 00:37:35 crc kubenswrapper[4953]: I1203 00:37:35.628090 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-interconnect-openstack-ca\" (UniqueName: \"kubernetes.io/secret/7e7fa27e-b88d-4d1c-b493-63b8f811dff5-default-interconnect-openstack-ca\") pod \"default-interconnect-68864d46cb-mv5rb\" (UID: \"7e7fa27e-b88d-4d1c-b493-63b8f811dff5\") " pod="service-telemetry/default-interconnect-68864d46cb-mv5rb" Dec 03 00:37:35 crc kubenswrapper[4953]: I1203 00:37:35.628119 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sasl-users\" (UniqueName: \"kubernetes.io/secret/7e7fa27e-b88d-4d1c-b493-63b8f811dff5-sasl-users\") pod \"default-interconnect-68864d46cb-mv5rb\" (UID: \"7e7fa27e-b88d-4d1c-b493-63b8f811dff5\") " pod="service-telemetry/default-interconnect-68864d46cb-mv5rb" Dec 03 00:37:35 crc kubenswrapper[4953]: I1203 00:37:35.729564 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-interconnect-inter-router-credentials\" (UniqueName: \"kubernetes.io/secret/7e7fa27e-b88d-4d1c-b493-63b8f811dff5-default-interconnect-inter-router-credentials\") pod \"default-interconnect-68864d46cb-mv5rb\" (UID: \"7e7fa27e-b88d-4d1c-b493-63b8f811dff5\") " pod="service-telemetry/default-interconnect-68864d46cb-mv5rb" Dec 03 00:37:35 crc kubenswrapper[4953]: I1203 00:37:35.729635 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-interconnect-openstack-credentials\" (UniqueName: \"kubernetes.io/secret/7e7fa27e-b88d-4d1c-b493-63b8f811dff5-default-interconnect-openstack-credentials\") pod \"default-interconnect-68864d46cb-mv5rb\" (UID: \"7e7fa27e-b88d-4d1c-b493-63b8f811dff5\") " pod="service-telemetry/default-interconnect-68864d46cb-mv5rb" Dec 03 00:37:35 crc kubenswrapper[4953]: I1203 00:37:35.729666 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-interconnect-inter-router-ca\" (UniqueName: \"kubernetes.io/secret/7e7fa27e-b88d-4d1c-b493-63b8f811dff5-default-interconnect-inter-router-ca\") pod \"default-interconnect-68864d46cb-mv5rb\" (UID: \"7e7fa27e-b88d-4d1c-b493-63b8f811dff5\") " pod="service-telemetry/default-interconnect-68864d46cb-mv5rb" Dec 03 00:37:35 crc kubenswrapper[4953]: I1203 00:37:35.729704 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sasl-config\" (UniqueName: \"kubernetes.io/configmap/7e7fa27e-b88d-4d1c-b493-63b8f811dff5-sasl-config\") pod \"default-interconnect-68864d46cb-mv5rb\" (UID: \"7e7fa27e-b88d-4d1c-b493-63b8f811dff5\") " pod="service-telemetry/default-interconnect-68864d46cb-mv5rb" Dec 03 00:37:35 crc kubenswrapper[4953]: I1203 00:37:35.729739 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-interconnect-openstack-ca\" (UniqueName: \"kubernetes.io/secret/7e7fa27e-b88d-4d1c-b493-63b8f811dff5-default-interconnect-openstack-ca\") pod \"default-interconnect-68864d46cb-mv5rb\" (UID: \"7e7fa27e-b88d-4d1c-b493-63b8f811dff5\") " pod="service-telemetry/default-interconnect-68864d46cb-mv5rb" Dec 03 00:37:35 crc kubenswrapper[4953]: I1203 00:37:35.729785 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sasl-users\" (UniqueName: \"kubernetes.io/secret/7e7fa27e-b88d-4d1c-b493-63b8f811dff5-sasl-users\") pod \"default-interconnect-68864d46cb-mv5rb\" (UID: \"7e7fa27e-b88d-4d1c-b493-63b8f811dff5\") " pod="service-telemetry/default-interconnect-68864d46cb-mv5rb" Dec 03 00:37:35 crc kubenswrapper[4953]: I1203 00:37:35.729826 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-shh7v\" (UniqueName: \"kubernetes.io/projected/7e7fa27e-b88d-4d1c-b493-63b8f811dff5-kube-api-access-shh7v\") pod \"default-interconnect-68864d46cb-mv5rb\" (UID: \"7e7fa27e-b88d-4d1c-b493-63b8f811dff5\") " pod="service-telemetry/default-interconnect-68864d46cb-mv5rb" Dec 03 00:37:35 crc kubenswrapper[4953]: I1203 00:37:35.731210 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sasl-config\" (UniqueName: \"kubernetes.io/configmap/7e7fa27e-b88d-4d1c-b493-63b8f811dff5-sasl-config\") pod \"default-interconnect-68864d46cb-mv5rb\" (UID: \"7e7fa27e-b88d-4d1c-b493-63b8f811dff5\") " pod="service-telemetry/default-interconnect-68864d46cb-mv5rb" Dec 03 00:37:35 crc kubenswrapper[4953]: I1203 00:37:35.741091 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sasl-users\" (UniqueName: \"kubernetes.io/secret/7e7fa27e-b88d-4d1c-b493-63b8f811dff5-sasl-users\") pod \"default-interconnect-68864d46cb-mv5rb\" (UID: \"7e7fa27e-b88d-4d1c-b493-63b8f811dff5\") " pod="service-telemetry/default-interconnect-68864d46cb-mv5rb" Dec 03 00:37:35 crc kubenswrapper[4953]: I1203 00:37:35.741189 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-interconnect-openstack-ca\" (UniqueName: \"kubernetes.io/secret/7e7fa27e-b88d-4d1c-b493-63b8f811dff5-default-interconnect-openstack-ca\") pod \"default-interconnect-68864d46cb-mv5rb\" (UID: \"7e7fa27e-b88d-4d1c-b493-63b8f811dff5\") " pod="service-telemetry/default-interconnect-68864d46cb-mv5rb" Dec 03 00:37:35 crc kubenswrapper[4953]: I1203 00:37:35.741609 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-interconnect-inter-router-ca\" (UniqueName: \"kubernetes.io/secret/7e7fa27e-b88d-4d1c-b493-63b8f811dff5-default-interconnect-inter-router-ca\") pod \"default-interconnect-68864d46cb-mv5rb\" (UID: \"7e7fa27e-b88d-4d1c-b493-63b8f811dff5\") " pod="service-telemetry/default-interconnect-68864d46cb-mv5rb" Dec 03 00:37:35 crc kubenswrapper[4953]: I1203 00:37:35.742302 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-interconnect-inter-router-credentials\" (UniqueName: \"kubernetes.io/secret/7e7fa27e-b88d-4d1c-b493-63b8f811dff5-default-interconnect-inter-router-credentials\") pod \"default-interconnect-68864d46cb-mv5rb\" (UID: \"7e7fa27e-b88d-4d1c-b493-63b8f811dff5\") " pod="service-telemetry/default-interconnect-68864d46cb-mv5rb" Dec 03 00:37:35 crc kubenswrapper[4953]: I1203 00:37:35.747782 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-interconnect-openstack-credentials\" (UniqueName: \"kubernetes.io/secret/7e7fa27e-b88d-4d1c-b493-63b8f811dff5-default-interconnect-openstack-credentials\") pod \"default-interconnect-68864d46cb-mv5rb\" (UID: \"7e7fa27e-b88d-4d1c-b493-63b8f811dff5\") " pod="service-telemetry/default-interconnect-68864d46cb-mv5rb" Dec 03 00:37:35 crc kubenswrapper[4953]: I1203 00:37:35.750708 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-shh7v\" (UniqueName: \"kubernetes.io/projected/7e7fa27e-b88d-4d1c-b493-63b8f811dff5-kube-api-access-shh7v\") pod \"default-interconnect-68864d46cb-mv5rb\" (UID: \"7e7fa27e-b88d-4d1c-b493-63b8f811dff5\") " pod="service-telemetry/default-interconnect-68864d46cb-mv5rb" Dec 03 00:37:35 crc kubenswrapper[4953]: I1203 00:37:35.909436 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/default-interconnect-68864d46cb-mv5rb" Dec 03 00:37:36 crc kubenswrapper[4953]: I1203 00:37:36.378222 4953 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/default-interconnect-68864d46cb-mv5rb"] Dec 03 00:37:36 crc kubenswrapper[4953]: I1203 00:37:36.461478 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-interconnect-68864d46cb-mv5rb" event={"ID":"7e7fa27e-b88d-4d1c-b493-63b8f811dff5","Type":"ContainerStarted","Data":"fb50727c588b8a1ca5306e94e7ef1c5437bfc069fe80d94177f2880f6b69023c"} Dec 03 00:37:41 crc kubenswrapper[4953]: I1203 00:37:41.366620 4953 scope.go:117] "RemoveContainer" containerID="150ffc325f215c6da77fc72304c13cb10dafc800bbc932afd319cc34442a3f76" Dec 03 00:37:41 crc kubenswrapper[4953]: E1203 00:37:41.369874 4953 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6dfbn_openshift-machine-config-operator(53ffb1e9-9a2e-4869-80d7-776f4aa7ef8b)\"" pod="openshift-machine-config-operator/machine-config-daemon-6dfbn" podUID="53ffb1e9-9a2e-4869-80d7-776f4aa7ef8b" Dec 03 00:37:42 crc kubenswrapper[4953]: I1203 00:37:42.513667 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-interconnect-68864d46cb-mv5rb" event={"ID":"7e7fa27e-b88d-4d1c-b493-63b8f811dff5","Type":"ContainerStarted","Data":"e5a15270b4aaa460870081d673e9a5c7305c0ac209e59b6e719624ffb24b521f"} Dec 03 00:37:42 crc kubenswrapper[4953]: I1203 00:37:42.566055 4953 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="service-telemetry/default-interconnect-68864d46cb-mv5rb" podStartSLOduration=1.855976138 podStartE2EDuration="7.5659604s" podCreationTimestamp="2025-12-03 00:37:35 +0000 UTC" firstStartedPulling="2025-12-03 00:37:36.387080237 +0000 UTC m=+1845.371361020" lastFinishedPulling="2025-12-03 00:37:42.097064499 +0000 UTC m=+1851.081345282" observedRunningTime="2025-12-03 00:37:42.534726348 +0000 UTC m=+1851.519007141" watchObservedRunningTime="2025-12-03 00:37:42.5659604 +0000 UTC m=+1851.550241183" Dec 03 00:37:47 crc kubenswrapper[4953]: I1203 00:37:47.076569 4953 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["service-telemetry/prometheus-default-0"] Dec 03 00:37:47 crc kubenswrapper[4953]: I1203 00:37:47.078716 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/prometheus-default-0" Dec 03 00:37:47 crc kubenswrapper[4953]: I1203 00:37:47.081147 4953 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"prometheus-default-web-config" Dec 03 00:37:47 crc kubenswrapper[4953]: I1203 00:37:47.081959 4953 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"default-session-secret" Dec 03 00:37:47 crc kubenswrapper[4953]: I1203 00:37:47.082381 4953 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"prometheus-default-tls-assets-0" Dec 03 00:37:47 crc kubenswrapper[4953]: I1203 00:37:47.082555 4953 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"default-prometheus-proxy-tls" Dec 03 00:37:47 crc kubenswrapper[4953]: I1203 00:37:47.082630 4953 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"prometheus-default" Dec 03 00:37:47 crc kubenswrapper[4953]: I1203 00:37:47.084200 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"serving-certs-ca-bundle" Dec 03 00:37:47 crc kubenswrapper[4953]: I1203 00:37:47.084343 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"prometheus-default-rulefiles-0" Dec 03 00:37:47 crc kubenswrapper[4953]: I1203 00:37:47.090610 4953 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"prometheus-stf-dockercfg-llfdt" Dec 03 00:37:47 crc kubenswrapper[4953]: I1203 00:37:47.101212 4953 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/prometheus-default-0"] Dec 03 00:37:47 crc kubenswrapper[4953]: I1203 00:37:47.144069 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/603c47f8-e2e2-4dbd-aa98-a0eca1acf697-tls-assets\") pod \"prometheus-default-0\" (UID: \"603c47f8-e2e2-4dbd-aa98-a0eca1acf697\") " pod="service-telemetry/prometheus-default-0" Dec 03 00:37:47 crc kubenswrapper[4953]: I1203 00:37:47.144146 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/603c47f8-e2e2-4dbd-aa98-a0eca1acf697-web-config\") pod \"prometheus-default-0\" (UID: \"603c47f8-e2e2-4dbd-aa98-a0eca1acf697\") " pod="service-telemetry/prometheus-default-0" Dec 03 00:37:47 crc kubenswrapper[4953]: I1203 00:37:47.144171 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/603c47f8-e2e2-4dbd-aa98-a0eca1acf697-config\") pod \"prometheus-default-0\" (UID: \"603c47f8-e2e2-4dbd-aa98-a0eca1acf697\") " pod="service-telemetry/prometheus-default-0" Dec 03 00:37:47 crc kubenswrapper[4953]: I1203 00:37:47.144232 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"configmap-serving-certs-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/603c47f8-e2e2-4dbd-aa98-a0eca1acf697-configmap-serving-certs-ca-bundle\") pod \"prometheus-default-0\" (UID: \"603c47f8-e2e2-4dbd-aa98-a0eca1acf697\") " pod="service-telemetry/prometheus-default-0" Dec 03 00:37:47 crc kubenswrapper[4953]: I1203 00:37:47.144294 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-d02c597a-c515-47b2-832e-9b9c9439bb47\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-d02c597a-c515-47b2-832e-9b9c9439bb47\") pod \"prometheus-default-0\" (UID: \"603c47f8-e2e2-4dbd-aa98-a0eca1acf697\") " pod="service-telemetry/prometheus-default-0" Dec 03 00:37:47 crc kubenswrapper[4953]: I1203 00:37:47.144384 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sdbl5\" (UniqueName: \"kubernetes.io/projected/603c47f8-e2e2-4dbd-aa98-a0eca1acf697-kube-api-access-sdbl5\") pod \"prometheus-default-0\" (UID: \"603c47f8-e2e2-4dbd-aa98-a0eca1acf697\") " pod="service-telemetry/prometheus-default-0" Dec 03 00:37:47 crc kubenswrapper[4953]: I1203 00:37:47.144820 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/603c47f8-e2e2-4dbd-aa98-a0eca1acf697-config-out\") pod \"prometheus-default-0\" (UID: \"603c47f8-e2e2-4dbd-aa98-a0eca1acf697\") " pod="service-telemetry/prometheus-default-0" Dec 03 00:37:47 crc kubenswrapper[4953]: I1203 00:37:47.144903 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-default-prometheus-proxy-tls\" (UniqueName: \"kubernetes.io/secret/603c47f8-e2e2-4dbd-aa98-a0eca1acf697-secret-default-prometheus-proxy-tls\") pod \"prometheus-default-0\" (UID: \"603c47f8-e2e2-4dbd-aa98-a0eca1acf697\") " pod="service-telemetry/prometheus-default-0" Dec 03 00:37:47 crc kubenswrapper[4953]: I1203 00:37:47.144939 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-default-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/603c47f8-e2e2-4dbd-aa98-a0eca1acf697-prometheus-default-rulefiles-0\") pod \"prometheus-default-0\" (UID: \"603c47f8-e2e2-4dbd-aa98-a0eca1acf697\") " pod="service-telemetry/prometheus-default-0" Dec 03 00:37:47 crc kubenswrapper[4953]: I1203 00:37:47.144990 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-default-session-secret\" (UniqueName: \"kubernetes.io/secret/603c47f8-e2e2-4dbd-aa98-a0eca1acf697-secret-default-session-secret\") pod \"prometheus-default-0\" (UID: \"603c47f8-e2e2-4dbd-aa98-a0eca1acf697\") " pod="service-telemetry/prometheus-default-0" Dec 03 00:37:47 crc kubenswrapper[4953]: I1203 00:37:47.247323 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-d02c597a-c515-47b2-832e-9b9c9439bb47\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-d02c597a-c515-47b2-832e-9b9c9439bb47\") pod \"prometheus-default-0\" (UID: \"603c47f8-e2e2-4dbd-aa98-a0eca1acf697\") " pod="service-telemetry/prometheus-default-0" Dec 03 00:37:47 crc kubenswrapper[4953]: I1203 00:37:47.247457 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sdbl5\" (UniqueName: \"kubernetes.io/projected/603c47f8-e2e2-4dbd-aa98-a0eca1acf697-kube-api-access-sdbl5\") pod \"prometheus-default-0\" (UID: \"603c47f8-e2e2-4dbd-aa98-a0eca1acf697\") " pod="service-telemetry/prometheus-default-0" Dec 03 00:37:47 crc kubenswrapper[4953]: I1203 00:37:47.247513 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/603c47f8-e2e2-4dbd-aa98-a0eca1acf697-config-out\") pod \"prometheus-default-0\" (UID: \"603c47f8-e2e2-4dbd-aa98-a0eca1acf697\") " pod="service-telemetry/prometheus-default-0" Dec 03 00:37:47 crc kubenswrapper[4953]: I1203 00:37:47.247568 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-default-prometheus-proxy-tls\" (UniqueName: \"kubernetes.io/secret/603c47f8-e2e2-4dbd-aa98-a0eca1acf697-secret-default-prometheus-proxy-tls\") pod \"prometheus-default-0\" (UID: \"603c47f8-e2e2-4dbd-aa98-a0eca1acf697\") " pod="service-telemetry/prometheus-default-0" Dec 03 00:37:47 crc kubenswrapper[4953]: I1203 00:37:47.247624 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"prometheus-default-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/603c47f8-e2e2-4dbd-aa98-a0eca1acf697-prometheus-default-rulefiles-0\") pod \"prometheus-default-0\" (UID: \"603c47f8-e2e2-4dbd-aa98-a0eca1acf697\") " pod="service-telemetry/prometheus-default-0" Dec 03 00:37:47 crc kubenswrapper[4953]: I1203 00:37:47.247872 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-default-session-secret\" (UniqueName: \"kubernetes.io/secret/603c47f8-e2e2-4dbd-aa98-a0eca1acf697-secret-default-session-secret\") pod \"prometheus-default-0\" (UID: \"603c47f8-e2e2-4dbd-aa98-a0eca1acf697\") " pod="service-telemetry/prometheus-default-0" Dec 03 00:37:47 crc kubenswrapper[4953]: E1203 00:37:47.247873 4953 secret.go:188] Couldn't get secret service-telemetry/default-prometheus-proxy-tls: secret "default-prometheus-proxy-tls" not found Dec 03 00:37:47 crc kubenswrapper[4953]: E1203 00:37:47.248156 4953 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/603c47f8-e2e2-4dbd-aa98-a0eca1acf697-secret-default-prometheus-proxy-tls podName:603c47f8-e2e2-4dbd-aa98-a0eca1acf697 nodeName:}" failed. No retries permitted until 2025-12-03 00:37:47.748122326 +0000 UTC m=+1856.732403109 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "secret-default-prometheus-proxy-tls" (UniqueName: "kubernetes.io/secret/603c47f8-e2e2-4dbd-aa98-a0eca1acf697-secret-default-prometheus-proxy-tls") pod "prometheus-default-0" (UID: "603c47f8-e2e2-4dbd-aa98-a0eca1acf697") : secret "default-prometheus-proxy-tls" not found Dec 03 00:37:47 crc kubenswrapper[4953]: I1203 00:37:47.249087 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"prometheus-default-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/603c47f8-e2e2-4dbd-aa98-a0eca1acf697-prometheus-default-rulefiles-0\") pod \"prometheus-default-0\" (UID: \"603c47f8-e2e2-4dbd-aa98-a0eca1acf697\") " pod="service-telemetry/prometheus-default-0" Dec 03 00:37:47 crc kubenswrapper[4953]: I1203 00:37:47.249201 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/603c47f8-e2e2-4dbd-aa98-a0eca1acf697-tls-assets\") pod \"prometheus-default-0\" (UID: \"603c47f8-e2e2-4dbd-aa98-a0eca1acf697\") " pod="service-telemetry/prometheus-default-0" Dec 03 00:37:47 crc kubenswrapper[4953]: I1203 00:37:47.249846 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/603c47f8-e2e2-4dbd-aa98-a0eca1acf697-web-config\") pod \"prometheus-default-0\" (UID: \"603c47f8-e2e2-4dbd-aa98-a0eca1acf697\") " pod="service-telemetry/prometheus-default-0" Dec 03 00:37:47 crc kubenswrapper[4953]: I1203 00:37:47.249896 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/603c47f8-e2e2-4dbd-aa98-a0eca1acf697-config\") pod \"prometheus-default-0\" (UID: \"603c47f8-e2e2-4dbd-aa98-a0eca1acf697\") " pod="service-telemetry/prometheus-default-0" Dec 03 00:37:47 crc kubenswrapper[4953]: I1203 00:37:47.250096 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"configmap-serving-certs-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/603c47f8-e2e2-4dbd-aa98-a0eca1acf697-configmap-serving-certs-ca-bundle\") pod \"prometheus-default-0\" (UID: \"603c47f8-e2e2-4dbd-aa98-a0eca1acf697\") " pod="service-telemetry/prometheus-default-0" Dec 03 00:37:47 crc kubenswrapper[4953]: I1203 00:37:47.251126 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"configmap-serving-certs-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/603c47f8-e2e2-4dbd-aa98-a0eca1acf697-configmap-serving-certs-ca-bundle\") pod \"prometheus-default-0\" (UID: \"603c47f8-e2e2-4dbd-aa98-a0eca1acf697\") " pod="service-telemetry/prometheus-default-0" Dec 03 00:37:47 crc kubenswrapper[4953]: I1203 00:37:47.252166 4953 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 03 00:37:47 crc kubenswrapper[4953]: I1203 00:37:47.252211 4953 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-d02c597a-c515-47b2-832e-9b9c9439bb47\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-d02c597a-c515-47b2-832e-9b9c9439bb47\") pod \"prometheus-default-0\" (UID: \"603c47f8-e2e2-4dbd-aa98-a0eca1acf697\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/effa09cd2414f77d0dd34cedf2661544f2a1fbeaccd60f9614b14ac0f031f04b/globalmount\"" pod="service-telemetry/prometheus-default-0" Dec 03 00:37:47 crc kubenswrapper[4953]: I1203 00:37:47.256250 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/603c47f8-e2e2-4dbd-aa98-a0eca1acf697-tls-assets\") pod \"prometheus-default-0\" (UID: \"603c47f8-e2e2-4dbd-aa98-a0eca1acf697\") " pod="service-telemetry/prometheus-default-0" Dec 03 00:37:47 crc kubenswrapper[4953]: I1203 00:37:47.258379 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/603c47f8-e2e2-4dbd-aa98-a0eca1acf697-config-out\") pod \"prometheus-default-0\" (UID: \"603c47f8-e2e2-4dbd-aa98-a0eca1acf697\") " pod="service-telemetry/prometheus-default-0" Dec 03 00:37:47 crc kubenswrapper[4953]: I1203 00:37:47.258474 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-default-session-secret\" (UniqueName: \"kubernetes.io/secret/603c47f8-e2e2-4dbd-aa98-a0eca1acf697-secret-default-session-secret\") pod \"prometheus-default-0\" (UID: \"603c47f8-e2e2-4dbd-aa98-a0eca1acf697\") " pod="service-telemetry/prometheus-default-0" Dec 03 00:37:47 crc kubenswrapper[4953]: I1203 00:37:47.258646 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/603c47f8-e2e2-4dbd-aa98-a0eca1acf697-config\") pod \"prometheus-default-0\" (UID: \"603c47f8-e2e2-4dbd-aa98-a0eca1acf697\") " pod="service-telemetry/prometheus-default-0" Dec 03 00:37:47 crc kubenswrapper[4953]: I1203 00:37:47.259724 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/603c47f8-e2e2-4dbd-aa98-a0eca1acf697-web-config\") pod \"prometheus-default-0\" (UID: \"603c47f8-e2e2-4dbd-aa98-a0eca1acf697\") " pod="service-telemetry/prometheus-default-0" Dec 03 00:37:47 crc kubenswrapper[4953]: I1203 00:37:47.268276 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sdbl5\" (UniqueName: \"kubernetes.io/projected/603c47f8-e2e2-4dbd-aa98-a0eca1acf697-kube-api-access-sdbl5\") pod \"prometheus-default-0\" (UID: \"603c47f8-e2e2-4dbd-aa98-a0eca1acf697\") " pod="service-telemetry/prometheus-default-0" Dec 03 00:37:47 crc kubenswrapper[4953]: I1203 00:37:47.297359 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-d02c597a-c515-47b2-832e-9b9c9439bb47\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-d02c597a-c515-47b2-832e-9b9c9439bb47\") pod \"prometheus-default-0\" (UID: \"603c47f8-e2e2-4dbd-aa98-a0eca1acf697\") " pod="service-telemetry/prometheus-default-0" Dec 03 00:37:47 crc kubenswrapper[4953]: I1203 00:37:47.760343 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-default-prometheus-proxy-tls\" (UniqueName: \"kubernetes.io/secret/603c47f8-e2e2-4dbd-aa98-a0eca1acf697-secret-default-prometheus-proxy-tls\") pod \"prometheus-default-0\" (UID: \"603c47f8-e2e2-4dbd-aa98-a0eca1acf697\") " pod="service-telemetry/prometheus-default-0" Dec 03 00:37:47 crc kubenswrapper[4953]: E1203 00:37:47.760693 4953 secret.go:188] Couldn't get secret service-telemetry/default-prometheus-proxy-tls: secret "default-prometheus-proxy-tls" not found Dec 03 00:37:47 crc kubenswrapper[4953]: E1203 00:37:47.760780 4953 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/603c47f8-e2e2-4dbd-aa98-a0eca1acf697-secret-default-prometheus-proxy-tls podName:603c47f8-e2e2-4dbd-aa98-a0eca1acf697 nodeName:}" failed. No retries permitted until 2025-12-03 00:37:48.760753658 +0000 UTC m=+1857.745034451 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "secret-default-prometheus-proxy-tls" (UniqueName: "kubernetes.io/secret/603c47f8-e2e2-4dbd-aa98-a0eca1acf697-secret-default-prometheus-proxy-tls") pod "prometheus-default-0" (UID: "603c47f8-e2e2-4dbd-aa98-a0eca1acf697") : secret "default-prometheus-proxy-tls" not found Dec 03 00:37:48 crc kubenswrapper[4953]: I1203 00:37:48.777281 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-default-prometheus-proxy-tls\" (UniqueName: \"kubernetes.io/secret/603c47f8-e2e2-4dbd-aa98-a0eca1acf697-secret-default-prometheus-proxy-tls\") pod \"prometheus-default-0\" (UID: \"603c47f8-e2e2-4dbd-aa98-a0eca1acf697\") " pod="service-telemetry/prometheus-default-0" Dec 03 00:37:48 crc kubenswrapper[4953]: I1203 00:37:48.783217 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-default-prometheus-proxy-tls\" (UniqueName: \"kubernetes.io/secret/603c47f8-e2e2-4dbd-aa98-a0eca1acf697-secret-default-prometheus-proxy-tls\") pod \"prometheus-default-0\" (UID: \"603c47f8-e2e2-4dbd-aa98-a0eca1acf697\") " pod="service-telemetry/prometheus-default-0" Dec 03 00:37:48 crc kubenswrapper[4953]: I1203 00:37:48.904775 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/prometheus-default-0" Dec 03 00:37:49 crc kubenswrapper[4953]: I1203 00:37:49.133840 4953 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/prometheus-default-0"] Dec 03 00:37:49 crc kubenswrapper[4953]: I1203 00:37:49.567504 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/prometheus-default-0" event={"ID":"603c47f8-e2e2-4dbd-aa98-a0eca1acf697","Type":"ContainerStarted","Data":"5f6911e15bb2fe9ed92c7131f4fa0dd195a071b6acfc21aeac3b312d3b1693bf"} Dec 03 00:37:52 crc kubenswrapper[4953]: I1203 00:37:52.348467 4953 scope.go:117] "RemoveContainer" containerID="150ffc325f215c6da77fc72304c13cb10dafc800bbc932afd319cc34442a3f76" Dec 03 00:37:52 crc kubenswrapper[4953]: E1203 00:37:52.349010 4953 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6dfbn_openshift-machine-config-operator(53ffb1e9-9a2e-4869-80d7-776f4aa7ef8b)\"" pod="openshift-machine-config-operator/machine-config-daemon-6dfbn" podUID="53ffb1e9-9a2e-4869-80d7-776f4aa7ef8b" Dec 03 00:37:54 crc kubenswrapper[4953]: I1203 00:37:54.611673 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/prometheus-default-0" event={"ID":"603c47f8-e2e2-4dbd-aa98-a0eca1acf697","Type":"ContainerStarted","Data":"2df2fcec770b950e6129c8bd22d6170670dc11a94ab8b532febcf8164cc37a25"} Dec 03 00:37:57 crc kubenswrapper[4953]: I1203 00:37:57.527598 4953 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["service-telemetry/default-snmp-webhook-6856cfb745-jf6rj"] Dec 03 00:37:57 crc kubenswrapper[4953]: I1203 00:37:57.529043 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/default-snmp-webhook-6856cfb745-jf6rj" Dec 03 00:37:57 crc kubenswrapper[4953]: I1203 00:37:57.541889 4953 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/default-snmp-webhook-6856cfb745-jf6rj"] Dec 03 00:37:57 crc kubenswrapper[4953]: I1203 00:37:57.627248 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-chlq9\" (UniqueName: \"kubernetes.io/projected/7bf37579-93eb-42d1-91ed-2ea71e4683ae-kube-api-access-chlq9\") pod \"default-snmp-webhook-6856cfb745-jf6rj\" (UID: \"7bf37579-93eb-42d1-91ed-2ea71e4683ae\") " pod="service-telemetry/default-snmp-webhook-6856cfb745-jf6rj" Dec 03 00:37:57 crc kubenswrapper[4953]: I1203 00:37:57.729440 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-chlq9\" (UniqueName: \"kubernetes.io/projected/7bf37579-93eb-42d1-91ed-2ea71e4683ae-kube-api-access-chlq9\") pod \"default-snmp-webhook-6856cfb745-jf6rj\" (UID: \"7bf37579-93eb-42d1-91ed-2ea71e4683ae\") " pod="service-telemetry/default-snmp-webhook-6856cfb745-jf6rj" Dec 03 00:37:57 crc kubenswrapper[4953]: I1203 00:37:57.754099 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-chlq9\" (UniqueName: \"kubernetes.io/projected/7bf37579-93eb-42d1-91ed-2ea71e4683ae-kube-api-access-chlq9\") pod \"default-snmp-webhook-6856cfb745-jf6rj\" (UID: \"7bf37579-93eb-42d1-91ed-2ea71e4683ae\") " pod="service-telemetry/default-snmp-webhook-6856cfb745-jf6rj" Dec 03 00:37:57 crc kubenswrapper[4953]: I1203 00:37:57.858763 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/default-snmp-webhook-6856cfb745-jf6rj" Dec 03 00:37:58 crc kubenswrapper[4953]: I1203 00:37:58.099353 4953 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/default-snmp-webhook-6856cfb745-jf6rj"] Dec 03 00:37:58 crc kubenswrapper[4953]: W1203 00:37:58.108264 4953 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7bf37579_93eb_42d1_91ed_2ea71e4683ae.slice/crio-9321552841fbb72f83af3bcf58fb81a0e064b3b9eca71b220f3c98edb181498f WatchSource:0}: Error finding container 9321552841fbb72f83af3bcf58fb81a0e064b3b9eca71b220f3c98edb181498f: Status 404 returned error can't find the container with id 9321552841fbb72f83af3bcf58fb81a0e064b3b9eca71b220f3c98edb181498f Dec 03 00:37:58 crc kubenswrapper[4953]: I1203 00:37:58.655616 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-snmp-webhook-6856cfb745-jf6rj" event={"ID":"7bf37579-93eb-42d1-91ed-2ea71e4683ae","Type":"ContainerStarted","Data":"9321552841fbb72f83af3bcf58fb81a0e064b3b9eca71b220f3c98edb181498f"} Dec 03 00:38:01 crc kubenswrapper[4953]: I1203 00:38:01.219607 4953 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["service-telemetry/alertmanager-default-0"] Dec 03 00:38:01 crc kubenswrapper[4953]: I1203 00:38:01.223186 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/alertmanager-default-0" Dec 03 00:38:01 crc kubenswrapper[4953]: I1203 00:38:01.230154 4953 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"alertmanager-default-cluster-tls-config" Dec 03 00:38:01 crc kubenswrapper[4953]: I1203 00:38:01.230228 4953 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"default-alertmanager-proxy-tls" Dec 03 00:38:01 crc kubenswrapper[4953]: I1203 00:38:01.230159 4953 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"alertmanager-default-generated" Dec 03 00:38:01 crc kubenswrapper[4953]: I1203 00:38:01.230159 4953 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"alertmanager-default-web-config" Dec 03 00:38:01 crc kubenswrapper[4953]: I1203 00:38:01.230612 4953 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"alertmanager-stf-dockercfg-dsxxq" Dec 03 00:38:01 crc kubenswrapper[4953]: I1203 00:38:01.240576 4953 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"alertmanager-default-tls-assets-0" Dec 03 00:38:01 crc kubenswrapper[4953]: I1203 00:38:01.252436 4953 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/alertmanager-default-0"] Dec 03 00:38:01 crc kubenswrapper[4953]: I1203 00:38:01.397640 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cluster-tls-config\" (UniqueName: \"kubernetes.io/secret/1df63057-30c5-4625-8365-9fe9907156c9-cluster-tls-config\") pod \"alertmanager-default-0\" (UID: \"1df63057-30c5-4625-8365-9fe9907156c9\") " pod="service-telemetry/alertmanager-default-0" Dec 03 00:38:01 crc kubenswrapper[4953]: I1203 00:38:01.397701 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-default-session-secret\" (UniqueName: \"kubernetes.io/secret/1df63057-30c5-4625-8365-9fe9907156c9-secret-default-session-secret\") pod \"alertmanager-default-0\" (UID: \"1df63057-30c5-4625-8365-9fe9907156c9\") " pod="service-telemetry/alertmanager-default-0" Dec 03 00:38:01 crc kubenswrapper[4953]: I1203 00:38:01.397730 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/1df63057-30c5-4625-8365-9fe9907156c9-web-config\") pod \"alertmanager-default-0\" (UID: \"1df63057-30c5-4625-8365-9fe9907156c9\") " pod="service-telemetry/alertmanager-default-0" Dec 03 00:38:01 crc kubenswrapper[4953]: I1203 00:38:01.397765 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/secret/1df63057-30c5-4625-8365-9fe9907156c9-config-volume\") pod \"alertmanager-default-0\" (UID: \"1df63057-30c5-4625-8365-9fe9907156c9\") " pod="service-telemetry/alertmanager-default-0" Dec 03 00:38:01 crc kubenswrapper[4953]: I1203 00:38:01.397790 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kf58b\" (UniqueName: \"kubernetes.io/projected/1df63057-30c5-4625-8365-9fe9907156c9-kube-api-access-kf58b\") pod \"alertmanager-default-0\" (UID: \"1df63057-30c5-4625-8365-9fe9907156c9\") " pod="service-telemetry/alertmanager-default-0" Dec 03 00:38:01 crc kubenswrapper[4953]: I1203 00:38:01.397812 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/1df63057-30c5-4625-8365-9fe9907156c9-tls-assets\") pod \"alertmanager-default-0\" (UID: \"1df63057-30c5-4625-8365-9fe9907156c9\") " pod="service-telemetry/alertmanager-default-0" Dec 03 00:38:01 crc kubenswrapper[4953]: I1203 00:38:01.397850 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/1df63057-30c5-4625-8365-9fe9907156c9-config-out\") pod \"alertmanager-default-0\" (UID: \"1df63057-30c5-4625-8365-9fe9907156c9\") " pod="service-telemetry/alertmanager-default-0" Dec 03 00:38:01 crc kubenswrapper[4953]: I1203 00:38:01.397879 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-default-alertmanager-proxy-tls\" (UniqueName: \"kubernetes.io/secret/1df63057-30c5-4625-8365-9fe9907156c9-secret-default-alertmanager-proxy-tls\") pod \"alertmanager-default-0\" (UID: \"1df63057-30c5-4625-8365-9fe9907156c9\") " pod="service-telemetry/alertmanager-default-0" Dec 03 00:38:01 crc kubenswrapper[4953]: I1203 00:38:01.397948 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-28ed638c-6802-4ffe-a59a-3bd9eec0ffc4\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-28ed638c-6802-4ffe-a59a-3bd9eec0ffc4\") pod \"alertmanager-default-0\" (UID: \"1df63057-30c5-4625-8365-9fe9907156c9\") " pod="service-telemetry/alertmanager-default-0" Dec 03 00:38:01 crc kubenswrapper[4953]: I1203 00:38:01.499265 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-default-alertmanager-proxy-tls\" (UniqueName: \"kubernetes.io/secret/1df63057-30c5-4625-8365-9fe9907156c9-secret-default-alertmanager-proxy-tls\") pod \"alertmanager-default-0\" (UID: \"1df63057-30c5-4625-8365-9fe9907156c9\") " pod="service-telemetry/alertmanager-default-0" Dec 03 00:38:01 crc kubenswrapper[4953]: E1203 00:38:01.499516 4953 secret.go:188] Couldn't get secret service-telemetry/default-alertmanager-proxy-tls: secret "default-alertmanager-proxy-tls" not found Dec 03 00:38:01 crc kubenswrapper[4953]: E1203 00:38:01.499615 4953 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/1df63057-30c5-4625-8365-9fe9907156c9-secret-default-alertmanager-proxy-tls podName:1df63057-30c5-4625-8365-9fe9907156c9 nodeName:}" failed. No retries permitted until 2025-12-03 00:38:01.999594626 +0000 UTC m=+1870.983875409 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "secret-default-alertmanager-proxy-tls" (UniqueName: "kubernetes.io/secret/1df63057-30c5-4625-8365-9fe9907156c9-secret-default-alertmanager-proxy-tls") pod "alertmanager-default-0" (UID: "1df63057-30c5-4625-8365-9fe9907156c9") : secret "default-alertmanager-proxy-tls" not found Dec 03 00:38:01 crc kubenswrapper[4953]: I1203 00:38:01.499835 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-28ed638c-6802-4ffe-a59a-3bd9eec0ffc4\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-28ed638c-6802-4ffe-a59a-3bd9eec0ffc4\") pod \"alertmanager-default-0\" (UID: \"1df63057-30c5-4625-8365-9fe9907156c9\") " pod="service-telemetry/alertmanager-default-0" Dec 03 00:38:01 crc kubenswrapper[4953]: I1203 00:38:01.500853 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cluster-tls-config\" (UniqueName: \"kubernetes.io/secret/1df63057-30c5-4625-8365-9fe9907156c9-cluster-tls-config\") pod \"alertmanager-default-0\" (UID: \"1df63057-30c5-4625-8365-9fe9907156c9\") " pod="service-telemetry/alertmanager-default-0" Dec 03 00:38:01 crc kubenswrapper[4953]: I1203 00:38:01.500883 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-default-session-secret\" (UniqueName: \"kubernetes.io/secret/1df63057-30c5-4625-8365-9fe9907156c9-secret-default-session-secret\") pod \"alertmanager-default-0\" (UID: \"1df63057-30c5-4625-8365-9fe9907156c9\") " pod="service-telemetry/alertmanager-default-0" Dec 03 00:38:01 crc kubenswrapper[4953]: I1203 00:38:01.500903 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/1df63057-30c5-4625-8365-9fe9907156c9-web-config\") pod \"alertmanager-default-0\" (UID: \"1df63057-30c5-4625-8365-9fe9907156c9\") " pod="service-telemetry/alertmanager-default-0" Dec 03 00:38:01 crc kubenswrapper[4953]: I1203 00:38:01.500941 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/secret/1df63057-30c5-4625-8365-9fe9907156c9-config-volume\") pod \"alertmanager-default-0\" (UID: \"1df63057-30c5-4625-8365-9fe9907156c9\") " pod="service-telemetry/alertmanager-default-0" Dec 03 00:38:01 crc kubenswrapper[4953]: I1203 00:38:01.500962 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kf58b\" (UniqueName: \"kubernetes.io/projected/1df63057-30c5-4625-8365-9fe9907156c9-kube-api-access-kf58b\") pod \"alertmanager-default-0\" (UID: \"1df63057-30c5-4625-8365-9fe9907156c9\") " pod="service-telemetry/alertmanager-default-0" Dec 03 00:38:01 crc kubenswrapper[4953]: I1203 00:38:01.500997 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/1df63057-30c5-4625-8365-9fe9907156c9-tls-assets\") pod \"alertmanager-default-0\" (UID: \"1df63057-30c5-4625-8365-9fe9907156c9\") " pod="service-telemetry/alertmanager-default-0" Dec 03 00:38:01 crc kubenswrapper[4953]: I1203 00:38:01.501040 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/1df63057-30c5-4625-8365-9fe9907156c9-config-out\") pod \"alertmanager-default-0\" (UID: \"1df63057-30c5-4625-8365-9fe9907156c9\") " pod="service-telemetry/alertmanager-default-0" Dec 03 00:38:01 crc kubenswrapper[4953]: I1203 00:38:01.505284 4953 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 03 00:38:01 crc kubenswrapper[4953]: I1203 00:38:01.505560 4953 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-28ed638c-6802-4ffe-a59a-3bd9eec0ffc4\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-28ed638c-6802-4ffe-a59a-3bd9eec0ffc4\") pod \"alertmanager-default-0\" (UID: \"1df63057-30c5-4625-8365-9fe9907156c9\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/8726b0e64f9dcf5e80fe8dc4104c40b9f40e26f3b1f2e33c2fc5545ba3b90f69/globalmount\"" pod="service-telemetry/alertmanager-default-0" Dec 03 00:38:01 crc kubenswrapper[4953]: I1203 00:38:01.511120 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-default-session-secret\" (UniqueName: \"kubernetes.io/secret/1df63057-30c5-4625-8365-9fe9907156c9-secret-default-session-secret\") pod \"alertmanager-default-0\" (UID: \"1df63057-30c5-4625-8365-9fe9907156c9\") " pod="service-telemetry/alertmanager-default-0" Dec 03 00:38:01 crc kubenswrapper[4953]: I1203 00:38:01.514157 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/secret/1df63057-30c5-4625-8365-9fe9907156c9-config-volume\") pod \"alertmanager-default-0\" (UID: \"1df63057-30c5-4625-8365-9fe9907156c9\") " pod="service-telemetry/alertmanager-default-0" Dec 03 00:38:01 crc kubenswrapper[4953]: I1203 00:38:01.515810 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/1df63057-30c5-4625-8365-9fe9907156c9-config-out\") pod \"alertmanager-default-0\" (UID: \"1df63057-30c5-4625-8365-9fe9907156c9\") " pod="service-telemetry/alertmanager-default-0" Dec 03 00:38:01 crc kubenswrapper[4953]: I1203 00:38:01.516582 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/1df63057-30c5-4625-8365-9fe9907156c9-web-config\") pod \"alertmanager-default-0\" (UID: \"1df63057-30c5-4625-8365-9fe9907156c9\") " pod="service-telemetry/alertmanager-default-0" Dec 03 00:38:01 crc kubenswrapper[4953]: I1203 00:38:01.520463 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/1df63057-30c5-4625-8365-9fe9907156c9-tls-assets\") pod \"alertmanager-default-0\" (UID: \"1df63057-30c5-4625-8365-9fe9907156c9\") " pod="service-telemetry/alertmanager-default-0" Dec 03 00:38:01 crc kubenswrapper[4953]: I1203 00:38:01.522731 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kf58b\" (UniqueName: \"kubernetes.io/projected/1df63057-30c5-4625-8365-9fe9907156c9-kube-api-access-kf58b\") pod \"alertmanager-default-0\" (UID: \"1df63057-30c5-4625-8365-9fe9907156c9\") " pod="service-telemetry/alertmanager-default-0" Dec 03 00:38:01 crc kubenswrapper[4953]: I1203 00:38:01.524848 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cluster-tls-config\" (UniqueName: \"kubernetes.io/secret/1df63057-30c5-4625-8365-9fe9907156c9-cluster-tls-config\") pod \"alertmanager-default-0\" (UID: \"1df63057-30c5-4625-8365-9fe9907156c9\") " pod="service-telemetry/alertmanager-default-0" Dec 03 00:38:01 crc kubenswrapper[4953]: I1203 00:38:01.563101 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-28ed638c-6802-4ffe-a59a-3bd9eec0ffc4\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-28ed638c-6802-4ffe-a59a-3bd9eec0ffc4\") pod \"alertmanager-default-0\" (UID: \"1df63057-30c5-4625-8365-9fe9907156c9\") " pod="service-telemetry/alertmanager-default-0" Dec 03 00:38:01 crc kubenswrapper[4953]: I1203 00:38:01.687855 4953 generic.go:334] "Generic (PLEG): container finished" podID="603c47f8-e2e2-4dbd-aa98-a0eca1acf697" containerID="2df2fcec770b950e6129c8bd22d6170670dc11a94ab8b532febcf8164cc37a25" exitCode=0 Dec 03 00:38:01 crc kubenswrapper[4953]: I1203 00:38:01.687905 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/prometheus-default-0" event={"ID":"603c47f8-e2e2-4dbd-aa98-a0eca1acf697","Type":"ContainerDied","Data":"2df2fcec770b950e6129c8bd22d6170670dc11a94ab8b532febcf8164cc37a25"} Dec 03 00:38:02 crc kubenswrapper[4953]: I1203 00:38:02.009787 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-default-alertmanager-proxy-tls\" (UniqueName: \"kubernetes.io/secret/1df63057-30c5-4625-8365-9fe9907156c9-secret-default-alertmanager-proxy-tls\") pod \"alertmanager-default-0\" (UID: \"1df63057-30c5-4625-8365-9fe9907156c9\") " pod="service-telemetry/alertmanager-default-0" Dec 03 00:38:02 crc kubenswrapper[4953]: E1203 00:38:02.010017 4953 secret.go:188] Couldn't get secret service-telemetry/default-alertmanager-proxy-tls: secret "default-alertmanager-proxy-tls" not found Dec 03 00:38:02 crc kubenswrapper[4953]: E1203 00:38:02.010107 4953 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/1df63057-30c5-4625-8365-9fe9907156c9-secret-default-alertmanager-proxy-tls podName:1df63057-30c5-4625-8365-9fe9907156c9 nodeName:}" failed. No retries permitted until 2025-12-03 00:38:03.010085285 +0000 UTC m=+1871.994366068 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "secret-default-alertmanager-proxy-tls" (UniqueName: "kubernetes.io/secret/1df63057-30c5-4625-8365-9fe9907156c9-secret-default-alertmanager-proxy-tls") pod "alertmanager-default-0" (UID: "1df63057-30c5-4625-8365-9fe9907156c9") : secret "default-alertmanager-proxy-tls" not found Dec 03 00:38:03 crc kubenswrapper[4953]: I1203 00:38:03.022959 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-default-alertmanager-proxy-tls\" (UniqueName: \"kubernetes.io/secret/1df63057-30c5-4625-8365-9fe9907156c9-secret-default-alertmanager-proxy-tls\") pod \"alertmanager-default-0\" (UID: \"1df63057-30c5-4625-8365-9fe9907156c9\") " pod="service-telemetry/alertmanager-default-0" Dec 03 00:38:03 crc kubenswrapper[4953]: E1203 00:38:03.023337 4953 secret.go:188] Couldn't get secret service-telemetry/default-alertmanager-proxy-tls: secret "default-alertmanager-proxy-tls" not found Dec 03 00:38:03 crc kubenswrapper[4953]: E1203 00:38:03.023417 4953 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/1df63057-30c5-4625-8365-9fe9907156c9-secret-default-alertmanager-proxy-tls podName:1df63057-30c5-4625-8365-9fe9907156c9 nodeName:}" failed. No retries permitted until 2025-12-03 00:38:05.023398354 +0000 UTC m=+1874.007679137 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "secret-default-alertmanager-proxy-tls" (UniqueName: "kubernetes.io/secret/1df63057-30c5-4625-8365-9fe9907156c9-secret-default-alertmanager-proxy-tls") pod "alertmanager-default-0" (UID: "1df63057-30c5-4625-8365-9fe9907156c9") : secret "default-alertmanager-proxy-tls" not found Dec 03 00:38:03 crc kubenswrapper[4953]: I1203 00:38:03.350141 4953 scope.go:117] "RemoveContainer" containerID="150ffc325f215c6da77fc72304c13cb10dafc800bbc932afd319cc34442a3f76" Dec 03 00:38:03 crc kubenswrapper[4953]: E1203 00:38:03.350450 4953 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6dfbn_openshift-machine-config-operator(53ffb1e9-9a2e-4869-80d7-776f4aa7ef8b)\"" pod="openshift-machine-config-operator/machine-config-daemon-6dfbn" podUID="53ffb1e9-9a2e-4869-80d7-776f4aa7ef8b" Dec 03 00:38:05 crc kubenswrapper[4953]: I1203 00:38:05.060582 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-default-alertmanager-proxy-tls\" (UniqueName: \"kubernetes.io/secret/1df63057-30c5-4625-8365-9fe9907156c9-secret-default-alertmanager-proxy-tls\") pod \"alertmanager-default-0\" (UID: \"1df63057-30c5-4625-8365-9fe9907156c9\") " pod="service-telemetry/alertmanager-default-0" Dec 03 00:38:05 crc kubenswrapper[4953]: I1203 00:38:05.071124 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-default-alertmanager-proxy-tls\" (UniqueName: \"kubernetes.io/secret/1df63057-30c5-4625-8365-9fe9907156c9-secret-default-alertmanager-proxy-tls\") pod \"alertmanager-default-0\" (UID: \"1df63057-30c5-4625-8365-9fe9907156c9\") " pod="service-telemetry/alertmanager-default-0" Dec 03 00:38:05 crc kubenswrapper[4953]: I1203 00:38:05.167873 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/alertmanager-default-0" Dec 03 00:38:09 crc kubenswrapper[4953]: I1203 00:38:09.489597 4953 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/alertmanager-default-0"] Dec 03 00:38:09 crc kubenswrapper[4953]: I1203 00:38:09.792179 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/alertmanager-default-0" event={"ID":"1df63057-30c5-4625-8365-9fe9907156c9","Type":"ContainerStarted","Data":"0edeb76a937f5d60495bbeda048d95e566fb200504e65b29098e8b9195972569"} Dec 03 00:38:10 crc kubenswrapper[4953]: I1203 00:38:10.802100 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-snmp-webhook-6856cfb745-jf6rj" event={"ID":"7bf37579-93eb-42d1-91ed-2ea71e4683ae","Type":"ContainerStarted","Data":"da8c6686d5ff3e37aa2bc3dd0b90d2ee7280a7133ae58cc094aa7d600de1da35"} Dec 03 00:38:10 crc kubenswrapper[4953]: I1203 00:38:10.830426 4953 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="service-telemetry/default-snmp-webhook-6856cfb745-jf6rj" podStartSLOduration=2.396803358 podStartE2EDuration="13.83039445s" podCreationTimestamp="2025-12-03 00:37:57 +0000 UTC" firstStartedPulling="2025-12-03 00:37:58.110783705 +0000 UTC m=+1867.095064478" lastFinishedPulling="2025-12-03 00:38:09.544374787 +0000 UTC m=+1878.528655570" observedRunningTime="2025-12-03 00:38:10.821751816 +0000 UTC m=+1879.806032599" watchObservedRunningTime="2025-12-03 00:38:10.83039445 +0000 UTC m=+1879.814675233" Dec 03 00:38:12 crc kubenswrapper[4953]: I1203 00:38:12.839033 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/alertmanager-default-0" event={"ID":"1df63057-30c5-4625-8365-9fe9907156c9","Type":"ContainerStarted","Data":"906cd88036eb34e71cfd7eedd023fd9df05cba9a27ff9795e1bb3e7a4f021c56"} Dec 03 00:38:15 crc kubenswrapper[4953]: I1203 00:38:15.871148 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/prometheus-default-0" event={"ID":"603c47f8-e2e2-4dbd-aa98-a0eca1acf697","Type":"ContainerStarted","Data":"d358e04b215aefc82cf76e50ae7fd3202f4f37e48538e476d2429099c0eea6a0"} Dec 03 00:38:16 crc kubenswrapper[4953]: I1203 00:38:16.348329 4953 scope.go:117] "RemoveContainer" containerID="150ffc325f215c6da77fc72304c13cb10dafc800bbc932afd319cc34442a3f76" Dec 03 00:38:16 crc kubenswrapper[4953]: E1203 00:38:16.349011 4953 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6dfbn_openshift-machine-config-operator(53ffb1e9-9a2e-4869-80d7-776f4aa7ef8b)\"" pod="openshift-machine-config-operator/machine-config-daemon-6dfbn" podUID="53ffb1e9-9a2e-4869-80d7-776f4aa7ef8b" Dec 03 00:38:17 crc kubenswrapper[4953]: I1203 00:38:17.888393 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/prometheus-default-0" event={"ID":"603c47f8-e2e2-4dbd-aa98-a0eca1acf697","Type":"ContainerStarted","Data":"fac222b7cbc1d890a32a43947736552552bda5f922c3a99bbc6516ab938627d6"} Dec 03 00:38:18 crc kubenswrapper[4953]: I1203 00:38:18.806175 4953 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["service-telemetry/default-cloud1-coll-meter-smartgateway-7cd87f9766-bgbxc"] Dec 03 00:38:18 crc kubenswrapper[4953]: I1203 00:38:18.809856 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/default-cloud1-coll-meter-smartgateway-7cd87f9766-bgbxc" Dec 03 00:38:18 crc kubenswrapper[4953]: I1203 00:38:18.814989 4953 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"smart-gateway-session-secret" Dec 03 00:38:18 crc kubenswrapper[4953]: I1203 00:38:18.815137 4953 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"smart-gateway-dockercfg-qqnq5" Dec 03 00:38:18 crc kubenswrapper[4953]: I1203 00:38:18.815329 4953 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"default-cloud1-coll-meter-proxy-tls" Dec 03 00:38:18 crc kubenswrapper[4953]: I1203 00:38:18.815445 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"default-cloud1-coll-meter-sg-core-configmap" Dec 03 00:38:18 crc kubenswrapper[4953]: I1203 00:38:18.821926 4953 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/default-cloud1-coll-meter-smartgateway-7cd87f9766-bgbxc"] Dec 03 00:38:18 crc kubenswrapper[4953]: I1203 00:38:18.850375 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7k86r\" (UniqueName: \"kubernetes.io/projected/f2ba70b7-7676-47af-aa77-4f90579e7ec1-kube-api-access-7k86r\") pod \"default-cloud1-coll-meter-smartgateway-7cd87f9766-bgbxc\" (UID: \"f2ba70b7-7676-47af-aa77-4f90579e7ec1\") " pod="service-telemetry/default-cloud1-coll-meter-smartgateway-7cd87f9766-bgbxc" Dec 03 00:38:18 crc kubenswrapper[4953]: I1203 00:38:18.850434 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"session-secret\" (UniqueName: \"kubernetes.io/secret/f2ba70b7-7676-47af-aa77-4f90579e7ec1-session-secret\") pod \"default-cloud1-coll-meter-smartgateway-7cd87f9766-bgbxc\" (UID: \"f2ba70b7-7676-47af-aa77-4f90579e7ec1\") " pod="service-telemetry/default-cloud1-coll-meter-smartgateway-7cd87f9766-bgbxc" Dec 03 00:38:18 crc kubenswrapper[4953]: I1203 00:38:18.850461 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-config\" (UniqueName: \"kubernetes.io/configmap/f2ba70b7-7676-47af-aa77-4f90579e7ec1-sg-core-config\") pod \"default-cloud1-coll-meter-smartgateway-7cd87f9766-bgbxc\" (UID: \"f2ba70b7-7676-47af-aa77-4f90579e7ec1\") " pod="service-telemetry/default-cloud1-coll-meter-smartgateway-7cd87f9766-bgbxc" Dec 03 00:38:18 crc kubenswrapper[4953]: I1203 00:38:18.850480 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/empty-dir/f2ba70b7-7676-47af-aa77-4f90579e7ec1-socket-dir\") pod \"default-cloud1-coll-meter-smartgateway-7cd87f9766-bgbxc\" (UID: \"f2ba70b7-7676-47af-aa77-4f90579e7ec1\") " pod="service-telemetry/default-cloud1-coll-meter-smartgateway-7cd87f9766-bgbxc" Dec 03 00:38:18 crc kubenswrapper[4953]: I1203 00:38:18.850503 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-cloud1-coll-meter-proxy-tls\" (UniqueName: \"kubernetes.io/secret/f2ba70b7-7676-47af-aa77-4f90579e7ec1-default-cloud1-coll-meter-proxy-tls\") pod \"default-cloud1-coll-meter-smartgateway-7cd87f9766-bgbxc\" (UID: \"f2ba70b7-7676-47af-aa77-4f90579e7ec1\") " pod="service-telemetry/default-cloud1-coll-meter-smartgateway-7cd87f9766-bgbxc" Dec 03 00:38:18 crc kubenswrapper[4953]: I1203 00:38:18.952495 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7k86r\" (UniqueName: \"kubernetes.io/projected/f2ba70b7-7676-47af-aa77-4f90579e7ec1-kube-api-access-7k86r\") pod \"default-cloud1-coll-meter-smartgateway-7cd87f9766-bgbxc\" (UID: \"f2ba70b7-7676-47af-aa77-4f90579e7ec1\") " pod="service-telemetry/default-cloud1-coll-meter-smartgateway-7cd87f9766-bgbxc" Dec 03 00:38:18 crc kubenswrapper[4953]: I1203 00:38:18.956216 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"session-secret\" (UniqueName: \"kubernetes.io/secret/f2ba70b7-7676-47af-aa77-4f90579e7ec1-session-secret\") pod \"default-cloud1-coll-meter-smartgateway-7cd87f9766-bgbxc\" (UID: \"f2ba70b7-7676-47af-aa77-4f90579e7ec1\") " pod="service-telemetry/default-cloud1-coll-meter-smartgateway-7cd87f9766-bgbxc" Dec 03 00:38:18 crc kubenswrapper[4953]: I1203 00:38:18.958261 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/empty-dir/f2ba70b7-7676-47af-aa77-4f90579e7ec1-socket-dir\") pod \"default-cloud1-coll-meter-smartgateway-7cd87f9766-bgbxc\" (UID: \"f2ba70b7-7676-47af-aa77-4f90579e7ec1\") " pod="service-telemetry/default-cloud1-coll-meter-smartgateway-7cd87f9766-bgbxc" Dec 03 00:38:18 crc kubenswrapper[4953]: I1203 00:38:18.958302 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-config\" (UniqueName: \"kubernetes.io/configmap/f2ba70b7-7676-47af-aa77-4f90579e7ec1-sg-core-config\") pod \"default-cloud1-coll-meter-smartgateway-7cd87f9766-bgbxc\" (UID: \"f2ba70b7-7676-47af-aa77-4f90579e7ec1\") " pod="service-telemetry/default-cloud1-coll-meter-smartgateway-7cd87f9766-bgbxc" Dec 03 00:38:18 crc kubenswrapper[4953]: I1203 00:38:18.958361 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-cloud1-coll-meter-proxy-tls\" (UniqueName: \"kubernetes.io/secret/f2ba70b7-7676-47af-aa77-4f90579e7ec1-default-cloud1-coll-meter-proxy-tls\") pod \"default-cloud1-coll-meter-smartgateway-7cd87f9766-bgbxc\" (UID: \"f2ba70b7-7676-47af-aa77-4f90579e7ec1\") " pod="service-telemetry/default-cloud1-coll-meter-smartgateway-7cd87f9766-bgbxc" Dec 03 00:38:18 crc kubenswrapper[4953]: E1203 00:38:18.958875 4953 secret.go:188] Couldn't get secret service-telemetry/default-cloud1-coll-meter-proxy-tls: secret "default-cloud1-coll-meter-proxy-tls" not found Dec 03 00:38:18 crc kubenswrapper[4953]: I1203 00:38:18.958985 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/empty-dir/f2ba70b7-7676-47af-aa77-4f90579e7ec1-socket-dir\") pod \"default-cloud1-coll-meter-smartgateway-7cd87f9766-bgbxc\" (UID: \"f2ba70b7-7676-47af-aa77-4f90579e7ec1\") " pod="service-telemetry/default-cloud1-coll-meter-smartgateway-7cd87f9766-bgbxc" Dec 03 00:38:18 crc kubenswrapper[4953]: E1203 00:38:18.959001 4953 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/f2ba70b7-7676-47af-aa77-4f90579e7ec1-default-cloud1-coll-meter-proxy-tls podName:f2ba70b7-7676-47af-aa77-4f90579e7ec1 nodeName:}" failed. No retries permitted until 2025-12-03 00:38:19.458927728 +0000 UTC m=+1888.443208511 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "default-cloud1-coll-meter-proxy-tls" (UniqueName: "kubernetes.io/secret/f2ba70b7-7676-47af-aa77-4f90579e7ec1-default-cloud1-coll-meter-proxy-tls") pod "default-cloud1-coll-meter-smartgateway-7cd87f9766-bgbxc" (UID: "f2ba70b7-7676-47af-aa77-4f90579e7ec1") : secret "default-cloud1-coll-meter-proxy-tls" not found Dec 03 00:38:18 crc kubenswrapper[4953]: I1203 00:38:18.960051 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-config\" (UniqueName: \"kubernetes.io/configmap/f2ba70b7-7676-47af-aa77-4f90579e7ec1-sg-core-config\") pod \"default-cloud1-coll-meter-smartgateway-7cd87f9766-bgbxc\" (UID: \"f2ba70b7-7676-47af-aa77-4f90579e7ec1\") " pod="service-telemetry/default-cloud1-coll-meter-smartgateway-7cd87f9766-bgbxc" Dec 03 00:38:18 crc kubenswrapper[4953]: I1203 00:38:18.964607 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"session-secret\" (UniqueName: \"kubernetes.io/secret/f2ba70b7-7676-47af-aa77-4f90579e7ec1-session-secret\") pod \"default-cloud1-coll-meter-smartgateway-7cd87f9766-bgbxc\" (UID: \"f2ba70b7-7676-47af-aa77-4f90579e7ec1\") " pod="service-telemetry/default-cloud1-coll-meter-smartgateway-7cd87f9766-bgbxc" Dec 03 00:38:18 crc kubenswrapper[4953]: I1203 00:38:18.973789 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7k86r\" (UniqueName: \"kubernetes.io/projected/f2ba70b7-7676-47af-aa77-4f90579e7ec1-kube-api-access-7k86r\") pod \"default-cloud1-coll-meter-smartgateway-7cd87f9766-bgbxc\" (UID: \"f2ba70b7-7676-47af-aa77-4f90579e7ec1\") " pod="service-telemetry/default-cloud1-coll-meter-smartgateway-7cd87f9766-bgbxc" Dec 03 00:38:19 crc kubenswrapper[4953]: I1203 00:38:19.467134 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-cloud1-coll-meter-proxy-tls\" (UniqueName: \"kubernetes.io/secret/f2ba70b7-7676-47af-aa77-4f90579e7ec1-default-cloud1-coll-meter-proxy-tls\") pod \"default-cloud1-coll-meter-smartgateway-7cd87f9766-bgbxc\" (UID: \"f2ba70b7-7676-47af-aa77-4f90579e7ec1\") " pod="service-telemetry/default-cloud1-coll-meter-smartgateway-7cd87f9766-bgbxc" Dec 03 00:38:19 crc kubenswrapper[4953]: E1203 00:38:19.467402 4953 secret.go:188] Couldn't get secret service-telemetry/default-cloud1-coll-meter-proxy-tls: secret "default-cloud1-coll-meter-proxy-tls" not found Dec 03 00:38:19 crc kubenswrapper[4953]: E1203 00:38:19.467550 4953 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/f2ba70b7-7676-47af-aa77-4f90579e7ec1-default-cloud1-coll-meter-proxy-tls podName:f2ba70b7-7676-47af-aa77-4f90579e7ec1 nodeName:}" failed. No retries permitted until 2025-12-03 00:38:20.467519766 +0000 UTC m=+1889.451800549 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "default-cloud1-coll-meter-proxy-tls" (UniqueName: "kubernetes.io/secret/f2ba70b7-7676-47af-aa77-4f90579e7ec1-default-cloud1-coll-meter-proxy-tls") pod "default-cloud1-coll-meter-smartgateway-7cd87f9766-bgbxc" (UID: "f2ba70b7-7676-47af-aa77-4f90579e7ec1") : secret "default-cloud1-coll-meter-proxy-tls" not found Dec 03 00:38:19 crc kubenswrapper[4953]: I1203 00:38:19.914881 4953 generic.go:334] "Generic (PLEG): container finished" podID="1df63057-30c5-4625-8365-9fe9907156c9" containerID="906cd88036eb34e71cfd7eedd023fd9df05cba9a27ff9795e1bb3e7a4f021c56" exitCode=0 Dec 03 00:38:19 crc kubenswrapper[4953]: I1203 00:38:19.915426 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/alertmanager-default-0" event={"ID":"1df63057-30c5-4625-8365-9fe9907156c9","Type":"ContainerDied","Data":"906cd88036eb34e71cfd7eedd023fd9df05cba9a27ff9795e1bb3e7a4f021c56"} Dec 03 00:38:20 crc kubenswrapper[4953]: I1203 00:38:20.485141 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-cloud1-coll-meter-proxy-tls\" (UniqueName: \"kubernetes.io/secret/f2ba70b7-7676-47af-aa77-4f90579e7ec1-default-cloud1-coll-meter-proxy-tls\") pod \"default-cloud1-coll-meter-smartgateway-7cd87f9766-bgbxc\" (UID: \"f2ba70b7-7676-47af-aa77-4f90579e7ec1\") " pod="service-telemetry/default-cloud1-coll-meter-smartgateway-7cd87f9766-bgbxc" Dec 03 00:38:20 crc kubenswrapper[4953]: I1203 00:38:20.493331 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-cloud1-coll-meter-proxy-tls\" (UniqueName: \"kubernetes.io/secret/f2ba70b7-7676-47af-aa77-4f90579e7ec1-default-cloud1-coll-meter-proxy-tls\") pod \"default-cloud1-coll-meter-smartgateway-7cd87f9766-bgbxc\" (UID: \"f2ba70b7-7676-47af-aa77-4f90579e7ec1\") " pod="service-telemetry/default-cloud1-coll-meter-smartgateway-7cd87f9766-bgbxc" Dec 03 00:38:20 crc kubenswrapper[4953]: I1203 00:38:20.646720 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/default-cloud1-coll-meter-smartgateway-7cd87f9766-bgbxc" Dec 03 00:38:21 crc kubenswrapper[4953]: I1203 00:38:21.227869 4953 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["service-telemetry/default-cloud1-ceil-meter-smartgateway-57948895dc-blgcn"] Dec 03 00:38:21 crc kubenswrapper[4953]: I1203 00:38:21.230920 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-57948895dc-blgcn" Dec 03 00:38:21 crc kubenswrapper[4953]: I1203 00:38:21.233734 4953 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"default-cloud1-ceil-meter-proxy-tls" Dec 03 00:38:21 crc kubenswrapper[4953]: I1203 00:38:21.233734 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"default-cloud1-ceil-meter-sg-core-configmap" Dec 03 00:38:21 crc kubenswrapper[4953]: I1203 00:38:21.264407 4953 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/default-cloud1-ceil-meter-smartgateway-57948895dc-blgcn"] Dec 03 00:38:21 crc kubenswrapper[4953]: I1203 00:38:21.400823 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-config\" (UniqueName: \"kubernetes.io/configmap/43d17ba9-6fdf-4dca-991c-924f350daa0b-sg-core-config\") pod \"default-cloud1-ceil-meter-smartgateway-57948895dc-blgcn\" (UID: \"43d17ba9-6fdf-4dca-991c-924f350daa0b\") " pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-57948895dc-blgcn" Dec 03 00:38:21 crc kubenswrapper[4953]: I1203 00:38:21.400903 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-cloud1-ceil-meter-proxy-tls\" (UniqueName: \"kubernetes.io/secret/43d17ba9-6fdf-4dca-991c-924f350daa0b-default-cloud1-ceil-meter-proxy-tls\") pod \"default-cloud1-ceil-meter-smartgateway-57948895dc-blgcn\" (UID: \"43d17ba9-6fdf-4dca-991c-924f350daa0b\") " pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-57948895dc-blgcn" Dec 03 00:38:21 crc kubenswrapper[4953]: I1203 00:38:21.400963 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/empty-dir/43d17ba9-6fdf-4dca-991c-924f350daa0b-socket-dir\") pod \"default-cloud1-ceil-meter-smartgateway-57948895dc-blgcn\" (UID: \"43d17ba9-6fdf-4dca-991c-924f350daa0b\") " pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-57948895dc-blgcn" Dec 03 00:38:21 crc kubenswrapper[4953]: I1203 00:38:21.401062 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"session-secret\" (UniqueName: \"kubernetes.io/secret/43d17ba9-6fdf-4dca-991c-924f350daa0b-session-secret\") pod \"default-cloud1-ceil-meter-smartgateway-57948895dc-blgcn\" (UID: \"43d17ba9-6fdf-4dca-991c-924f350daa0b\") " pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-57948895dc-blgcn" Dec 03 00:38:21 crc kubenswrapper[4953]: I1203 00:38:21.401090 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4xn4p\" (UniqueName: \"kubernetes.io/projected/43d17ba9-6fdf-4dca-991c-924f350daa0b-kube-api-access-4xn4p\") pod \"default-cloud1-ceil-meter-smartgateway-57948895dc-blgcn\" (UID: \"43d17ba9-6fdf-4dca-991c-924f350daa0b\") " pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-57948895dc-blgcn" Dec 03 00:38:21 crc kubenswrapper[4953]: I1203 00:38:21.503484 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/empty-dir/43d17ba9-6fdf-4dca-991c-924f350daa0b-socket-dir\") pod \"default-cloud1-ceil-meter-smartgateway-57948895dc-blgcn\" (UID: \"43d17ba9-6fdf-4dca-991c-924f350daa0b\") " pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-57948895dc-blgcn" Dec 03 00:38:21 crc kubenswrapper[4953]: I1203 00:38:21.503638 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"session-secret\" (UniqueName: \"kubernetes.io/secret/43d17ba9-6fdf-4dca-991c-924f350daa0b-session-secret\") pod \"default-cloud1-ceil-meter-smartgateway-57948895dc-blgcn\" (UID: \"43d17ba9-6fdf-4dca-991c-924f350daa0b\") " pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-57948895dc-blgcn" Dec 03 00:38:21 crc kubenswrapper[4953]: I1203 00:38:21.503668 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4xn4p\" (UniqueName: \"kubernetes.io/projected/43d17ba9-6fdf-4dca-991c-924f350daa0b-kube-api-access-4xn4p\") pod \"default-cloud1-ceil-meter-smartgateway-57948895dc-blgcn\" (UID: \"43d17ba9-6fdf-4dca-991c-924f350daa0b\") " pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-57948895dc-blgcn" Dec 03 00:38:21 crc kubenswrapper[4953]: I1203 00:38:21.503729 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-config\" (UniqueName: \"kubernetes.io/configmap/43d17ba9-6fdf-4dca-991c-924f350daa0b-sg-core-config\") pod \"default-cloud1-ceil-meter-smartgateway-57948895dc-blgcn\" (UID: \"43d17ba9-6fdf-4dca-991c-924f350daa0b\") " pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-57948895dc-blgcn" Dec 03 00:38:21 crc kubenswrapper[4953]: I1203 00:38:21.503782 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-cloud1-ceil-meter-proxy-tls\" (UniqueName: \"kubernetes.io/secret/43d17ba9-6fdf-4dca-991c-924f350daa0b-default-cloud1-ceil-meter-proxy-tls\") pod \"default-cloud1-ceil-meter-smartgateway-57948895dc-blgcn\" (UID: \"43d17ba9-6fdf-4dca-991c-924f350daa0b\") " pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-57948895dc-blgcn" Dec 03 00:38:21 crc kubenswrapper[4953]: E1203 00:38:21.504049 4953 secret.go:188] Couldn't get secret service-telemetry/default-cloud1-ceil-meter-proxy-tls: secret "default-cloud1-ceil-meter-proxy-tls" not found Dec 03 00:38:21 crc kubenswrapper[4953]: E1203 00:38:21.504149 4953 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/43d17ba9-6fdf-4dca-991c-924f350daa0b-default-cloud1-ceil-meter-proxy-tls podName:43d17ba9-6fdf-4dca-991c-924f350daa0b nodeName:}" failed. No retries permitted until 2025-12-03 00:38:22.004116847 +0000 UTC m=+1890.988397630 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "default-cloud1-ceil-meter-proxy-tls" (UniqueName: "kubernetes.io/secret/43d17ba9-6fdf-4dca-991c-924f350daa0b-default-cloud1-ceil-meter-proxy-tls") pod "default-cloud1-ceil-meter-smartgateway-57948895dc-blgcn" (UID: "43d17ba9-6fdf-4dca-991c-924f350daa0b") : secret "default-cloud1-ceil-meter-proxy-tls" not found Dec 03 00:38:21 crc kubenswrapper[4953]: I1203 00:38:21.504314 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/empty-dir/43d17ba9-6fdf-4dca-991c-924f350daa0b-socket-dir\") pod \"default-cloud1-ceil-meter-smartgateway-57948895dc-blgcn\" (UID: \"43d17ba9-6fdf-4dca-991c-924f350daa0b\") " pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-57948895dc-blgcn" Dec 03 00:38:21 crc kubenswrapper[4953]: I1203 00:38:21.505054 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-config\" (UniqueName: \"kubernetes.io/configmap/43d17ba9-6fdf-4dca-991c-924f350daa0b-sg-core-config\") pod \"default-cloud1-ceil-meter-smartgateway-57948895dc-blgcn\" (UID: \"43d17ba9-6fdf-4dca-991c-924f350daa0b\") " pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-57948895dc-blgcn" Dec 03 00:38:21 crc kubenswrapper[4953]: I1203 00:38:21.523175 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4xn4p\" (UniqueName: \"kubernetes.io/projected/43d17ba9-6fdf-4dca-991c-924f350daa0b-kube-api-access-4xn4p\") pod \"default-cloud1-ceil-meter-smartgateway-57948895dc-blgcn\" (UID: \"43d17ba9-6fdf-4dca-991c-924f350daa0b\") " pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-57948895dc-blgcn" Dec 03 00:38:21 crc kubenswrapper[4953]: I1203 00:38:21.528553 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"session-secret\" (UniqueName: \"kubernetes.io/secret/43d17ba9-6fdf-4dca-991c-924f350daa0b-session-secret\") pod \"default-cloud1-ceil-meter-smartgateway-57948895dc-blgcn\" (UID: \"43d17ba9-6fdf-4dca-991c-924f350daa0b\") " pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-57948895dc-blgcn" Dec 03 00:38:22 crc kubenswrapper[4953]: I1203 00:38:22.012523 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-cloud1-ceil-meter-proxy-tls\" (UniqueName: \"kubernetes.io/secret/43d17ba9-6fdf-4dca-991c-924f350daa0b-default-cloud1-ceil-meter-proxy-tls\") pod \"default-cloud1-ceil-meter-smartgateway-57948895dc-blgcn\" (UID: \"43d17ba9-6fdf-4dca-991c-924f350daa0b\") " pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-57948895dc-blgcn" Dec 03 00:38:22 crc kubenswrapper[4953]: E1203 00:38:22.012783 4953 secret.go:188] Couldn't get secret service-telemetry/default-cloud1-ceil-meter-proxy-tls: secret "default-cloud1-ceil-meter-proxy-tls" not found Dec 03 00:38:22 crc kubenswrapper[4953]: E1203 00:38:22.012911 4953 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/43d17ba9-6fdf-4dca-991c-924f350daa0b-default-cloud1-ceil-meter-proxy-tls podName:43d17ba9-6fdf-4dca-991c-924f350daa0b nodeName:}" failed. No retries permitted until 2025-12-03 00:38:23.012877339 +0000 UTC m=+1891.997158162 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "default-cloud1-ceil-meter-proxy-tls" (UniqueName: "kubernetes.io/secret/43d17ba9-6fdf-4dca-991c-924f350daa0b-default-cloud1-ceil-meter-proxy-tls") pod "default-cloud1-ceil-meter-smartgateway-57948895dc-blgcn" (UID: "43d17ba9-6fdf-4dca-991c-924f350daa0b") : secret "default-cloud1-ceil-meter-proxy-tls" not found Dec 03 00:38:23 crc kubenswrapper[4953]: I1203 00:38:23.029150 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-cloud1-ceil-meter-proxy-tls\" (UniqueName: \"kubernetes.io/secret/43d17ba9-6fdf-4dca-991c-924f350daa0b-default-cloud1-ceil-meter-proxy-tls\") pod \"default-cloud1-ceil-meter-smartgateway-57948895dc-blgcn\" (UID: \"43d17ba9-6fdf-4dca-991c-924f350daa0b\") " pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-57948895dc-blgcn" Dec 03 00:38:23 crc kubenswrapper[4953]: I1203 00:38:23.049326 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-cloud1-ceil-meter-proxy-tls\" (UniqueName: \"kubernetes.io/secret/43d17ba9-6fdf-4dca-991c-924f350daa0b-default-cloud1-ceil-meter-proxy-tls\") pod \"default-cloud1-ceil-meter-smartgateway-57948895dc-blgcn\" (UID: \"43d17ba9-6fdf-4dca-991c-924f350daa0b\") " pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-57948895dc-blgcn" Dec 03 00:38:23 crc kubenswrapper[4953]: I1203 00:38:23.087761 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-57948895dc-blgcn" Dec 03 00:38:24 crc kubenswrapper[4953]: I1203 00:38:24.714321 4953 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/default-cloud1-ceil-meter-smartgateway-57948895dc-blgcn"] Dec 03 00:38:25 crc kubenswrapper[4953]: I1203 00:38:25.609135 4953 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["service-telemetry/default-cloud1-sens-meter-smartgateway-5759b4d97-nlgdm"] Dec 03 00:38:25 crc kubenswrapper[4953]: I1203 00:38:25.612032 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/default-cloud1-sens-meter-smartgateway-5759b4d97-nlgdm" Dec 03 00:38:25 crc kubenswrapper[4953]: I1203 00:38:25.620252 4953 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"default-cloud1-sens-meter-proxy-tls" Dec 03 00:38:25 crc kubenswrapper[4953]: I1203 00:38:25.620272 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"default-cloud1-sens-meter-sg-core-configmap" Dec 03 00:38:25 crc kubenswrapper[4953]: I1203 00:38:25.630211 4953 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/default-cloud1-sens-meter-smartgateway-5759b4d97-nlgdm"] Dec 03 00:38:25 crc kubenswrapper[4953]: I1203 00:38:25.779617 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"session-secret\" (UniqueName: \"kubernetes.io/secret/d660753e-48eb-4388-bf97-c7050b9cb1d2-session-secret\") pod \"default-cloud1-sens-meter-smartgateway-5759b4d97-nlgdm\" (UID: \"d660753e-48eb-4388-bf97-c7050b9cb1d2\") " pod="service-telemetry/default-cloud1-sens-meter-smartgateway-5759b4d97-nlgdm" Dec 03 00:38:25 crc kubenswrapper[4953]: I1203 00:38:25.780361 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-cloud1-sens-meter-proxy-tls\" (UniqueName: \"kubernetes.io/secret/d660753e-48eb-4388-bf97-c7050b9cb1d2-default-cloud1-sens-meter-proxy-tls\") pod \"default-cloud1-sens-meter-smartgateway-5759b4d97-nlgdm\" (UID: \"d660753e-48eb-4388-bf97-c7050b9cb1d2\") " pod="service-telemetry/default-cloud1-sens-meter-smartgateway-5759b4d97-nlgdm" Dec 03 00:38:25 crc kubenswrapper[4953]: I1203 00:38:25.780617 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qclmr\" (UniqueName: \"kubernetes.io/projected/d660753e-48eb-4388-bf97-c7050b9cb1d2-kube-api-access-qclmr\") pod \"default-cloud1-sens-meter-smartgateway-5759b4d97-nlgdm\" (UID: \"d660753e-48eb-4388-bf97-c7050b9cb1d2\") " pod="service-telemetry/default-cloud1-sens-meter-smartgateway-5759b4d97-nlgdm" Dec 03 00:38:25 crc kubenswrapper[4953]: I1203 00:38:25.780691 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/empty-dir/d660753e-48eb-4388-bf97-c7050b9cb1d2-socket-dir\") pod \"default-cloud1-sens-meter-smartgateway-5759b4d97-nlgdm\" (UID: \"d660753e-48eb-4388-bf97-c7050b9cb1d2\") " pod="service-telemetry/default-cloud1-sens-meter-smartgateway-5759b4d97-nlgdm" Dec 03 00:38:25 crc kubenswrapper[4953]: I1203 00:38:25.780811 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-config\" (UniqueName: \"kubernetes.io/configmap/d660753e-48eb-4388-bf97-c7050b9cb1d2-sg-core-config\") pod \"default-cloud1-sens-meter-smartgateway-5759b4d97-nlgdm\" (UID: \"d660753e-48eb-4388-bf97-c7050b9cb1d2\") " pod="service-telemetry/default-cloud1-sens-meter-smartgateway-5759b4d97-nlgdm" Dec 03 00:38:25 crc kubenswrapper[4953]: I1203 00:38:25.882989 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"session-secret\" (UniqueName: \"kubernetes.io/secret/d660753e-48eb-4388-bf97-c7050b9cb1d2-session-secret\") pod \"default-cloud1-sens-meter-smartgateway-5759b4d97-nlgdm\" (UID: \"d660753e-48eb-4388-bf97-c7050b9cb1d2\") " pod="service-telemetry/default-cloud1-sens-meter-smartgateway-5759b4d97-nlgdm" Dec 03 00:38:25 crc kubenswrapper[4953]: I1203 00:38:25.883070 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-cloud1-sens-meter-proxy-tls\" (UniqueName: \"kubernetes.io/secret/d660753e-48eb-4388-bf97-c7050b9cb1d2-default-cloud1-sens-meter-proxy-tls\") pod \"default-cloud1-sens-meter-smartgateway-5759b4d97-nlgdm\" (UID: \"d660753e-48eb-4388-bf97-c7050b9cb1d2\") " pod="service-telemetry/default-cloud1-sens-meter-smartgateway-5759b4d97-nlgdm" Dec 03 00:38:25 crc kubenswrapper[4953]: I1203 00:38:25.883145 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qclmr\" (UniqueName: \"kubernetes.io/projected/d660753e-48eb-4388-bf97-c7050b9cb1d2-kube-api-access-qclmr\") pod \"default-cloud1-sens-meter-smartgateway-5759b4d97-nlgdm\" (UID: \"d660753e-48eb-4388-bf97-c7050b9cb1d2\") " pod="service-telemetry/default-cloud1-sens-meter-smartgateway-5759b4d97-nlgdm" Dec 03 00:38:25 crc kubenswrapper[4953]: I1203 00:38:25.883189 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/empty-dir/d660753e-48eb-4388-bf97-c7050b9cb1d2-socket-dir\") pod \"default-cloud1-sens-meter-smartgateway-5759b4d97-nlgdm\" (UID: \"d660753e-48eb-4388-bf97-c7050b9cb1d2\") " pod="service-telemetry/default-cloud1-sens-meter-smartgateway-5759b4d97-nlgdm" Dec 03 00:38:25 crc kubenswrapper[4953]: I1203 00:38:25.883235 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-config\" (UniqueName: \"kubernetes.io/configmap/d660753e-48eb-4388-bf97-c7050b9cb1d2-sg-core-config\") pod \"default-cloud1-sens-meter-smartgateway-5759b4d97-nlgdm\" (UID: \"d660753e-48eb-4388-bf97-c7050b9cb1d2\") " pod="service-telemetry/default-cloud1-sens-meter-smartgateway-5759b4d97-nlgdm" Dec 03 00:38:25 crc kubenswrapper[4953]: E1203 00:38:25.883266 4953 secret.go:188] Couldn't get secret service-telemetry/default-cloud1-sens-meter-proxy-tls: secret "default-cloud1-sens-meter-proxy-tls" not found Dec 03 00:38:25 crc kubenswrapper[4953]: E1203 00:38:25.883356 4953 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/d660753e-48eb-4388-bf97-c7050b9cb1d2-default-cloud1-sens-meter-proxy-tls podName:d660753e-48eb-4388-bf97-c7050b9cb1d2 nodeName:}" failed. No retries permitted until 2025-12-03 00:38:26.383334891 +0000 UTC m=+1895.367615674 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "default-cloud1-sens-meter-proxy-tls" (UniqueName: "kubernetes.io/secret/d660753e-48eb-4388-bf97-c7050b9cb1d2-default-cloud1-sens-meter-proxy-tls") pod "default-cloud1-sens-meter-smartgateway-5759b4d97-nlgdm" (UID: "d660753e-48eb-4388-bf97-c7050b9cb1d2") : secret "default-cloud1-sens-meter-proxy-tls" not found Dec 03 00:38:25 crc kubenswrapper[4953]: I1203 00:38:25.883856 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/empty-dir/d660753e-48eb-4388-bf97-c7050b9cb1d2-socket-dir\") pod \"default-cloud1-sens-meter-smartgateway-5759b4d97-nlgdm\" (UID: \"d660753e-48eb-4388-bf97-c7050b9cb1d2\") " pod="service-telemetry/default-cloud1-sens-meter-smartgateway-5759b4d97-nlgdm" Dec 03 00:38:25 crc kubenswrapper[4953]: I1203 00:38:25.885605 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-config\" (UniqueName: \"kubernetes.io/configmap/d660753e-48eb-4388-bf97-c7050b9cb1d2-sg-core-config\") pod \"default-cloud1-sens-meter-smartgateway-5759b4d97-nlgdm\" (UID: \"d660753e-48eb-4388-bf97-c7050b9cb1d2\") " pod="service-telemetry/default-cloud1-sens-meter-smartgateway-5759b4d97-nlgdm" Dec 03 00:38:25 crc kubenswrapper[4953]: I1203 00:38:25.892709 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"session-secret\" (UniqueName: \"kubernetes.io/secret/d660753e-48eb-4388-bf97-c7050b9cb1d2-session-secret\") pod \"default-cloud1-sens-meter-smartgateway-5759b4d97-nlgdm\" (UID: \"d660753e-48eb-4388-bf97-c7050b9cb1d2\") " pod="service-telemetry/default-cloud1-sens-meter-smartgateway-5759b4d97-nlgdm" Dec 03 00:38:25 crc kubenswrapper[4953]: I1203 00:38:25.908966 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qclmr\" (UniqueName: \"kubernetes.io/projected/d660753e-48eb-4388-bf97-c7050b9cb1d2-kube-api-access-qclmr\") pod \"default-cloud1-sens-meter-smartgateway-5759b4d97-nlgdm\" (UID: \"d660753e-48eb-4388-bf97-c7050b9cb1d2\") " pod="service-telemetry/default-cloud1-sens-meter-smartgateway-5759b4d97-nlgdm" Dec 03 00:38:25 crc kubenswrapper[4953]: I1203 00:38:25.965381 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-57948895dc-blgcn" event={"ID":"43d17ba9-6fdf-4dca-991c-924f350daa0b","Type":"ContainerStarted","Data":"f14bf8ddc9891268b56eedfd608913cc2f4fa71146c590e47a7986d36e09112c"} Dec 03 00:38:26 crc kubenswrapper[4953]: I1203 00:38:26.390083 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-cloud1-sens-meter-proxy-tls\" (UniqueName: \"kubernetes.io/secret/d660753e-48eb-4388-bf97-c7050b9cb1d2-default-cloud1-sens-meter-proxy-tls\") pod \"default-cloud1-sens-meter-smartgateway-5759b4d97-nlgdm\" (UID: \"d660753e-48eb-4388-bf97-c7050b9cb1d2\") " pod="service-telemetry/default-cloud1-sens-meter-smartgateway-5759b4d97-nlgdm" Dec 03 00:38:26 crc kubenswrapper[4953]: E1203 00:38:26.390781 4953 secret.go:188] Couldn't get secret service-telemetry/default-cloud1-sens-meter-proxy-tls: secret "default-cloud1-sens-meter-proxy-tls" not found Dec 03 00:38:26 crc kubenswrapper[4953]: E1203 00:38:26.390845 4953 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/d660753e-48eb-4388-bf97-c7050b9cb1d2-default-cloud1-sens-meter-proxy-tls podName:d660753e-48eb-4388-bf97-c7050b9cb1d2 nodeName:}" failed. No retries permitted until 2025-12-03 00:38:27.390827083 +0000 UTC m=+1896.375107866 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "default-cloud1-sens-meter-proxy-tls" (UniqueName: "kubernetes.io/secret/d660753e-48eb-4388-bf97-c7050b9cb1d2-default-cloud1-sens-meter-proxy-tls") pod "default-cloud1-sens-meter-smartgateway-5759b4d97-nlgdm" (UID: "d660753e-48eb-4388-bf97-c7050b9cb1d2") : secret "default-cloud1-sens-meter-proxy-tls" not found Dec 03 00:38:27 crc kubenswrapper[4953]: I1203 00:38:27.166557 4953 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/default-cloud1-coll-meter-smartgateway-7cd87f9766-bgbxc"] Dec 03 00:38:27 crc kubenswrapper[4953]: I1203 00:38:27.409484 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-cloud1-sens-meter-proxy-tls\" (UniqueName: \"kubernetes.io/secret/d660753e-48eb-4388-bf97-c7050b9cb1d2-default-cloud1-sens-meter-proxy-tls\") pod \"default-cloud1-sens-meter-smartgateway-5759b4d97-nlgdm\" (UID: \"d660753e-48eb-4388-bf97-c7050b9cb1d2\") " pod="service-telemetry/default-cloud1-sens-meter-smartgateway-5759b4d97-nlgdm" Dec 03 00:38:27 crc kubenswrapper[4953]: I1203 00:38:27.419873 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-cloud1-sens-meter-proxy-tls\" (UniqueName: \"kubernetes.io/secret/d660753e-48eb-4388-bf97-c7050b9cb1d2-default-cloud1-sens-meter-proxy-tls\") pod \"default-cloud1-sens-meter-smartgateway-5759b4d97-nlgdm\" (UID: \"d660753e-48eb-4388-bf97-c7050b9cb1d2\") " pod="service-telemetry/default-cloud1-sens-meter-smartgateway-5759b4d97-nlgdm" Dec 03 00:38:27 crc kubenswrapper[4953]: I1203 00:38:27.471196 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/default-cloud1-sens-meter-smartgateway-5759b4d97-nlgdm" Dec 03 00:38:27 crc kubenswrapper[4953]: I1203 00:38:27.854108 4953 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/default-cloud1-sens-meter-smartgateway-5759b4d97-nlgdm"] Dec 03 00:38:28 crc kubenswrapper[4953]: I1203 00:38:28.003095 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-57948895dc-blgcn" event={"ID":"43d17ba9-6fdf-4dca-991c-924f350daa0b","Type":"ContainerStarted","Data":"d628218047cd0a74b37283f3ecfbce1c46b8b122b5bb15959982206be1fa05f5"} Dec 03 00:38:28 crc kubenswrapper[4953]: I1203 00:38:28.006295 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/alertmanager-default-0" event={"ID":"1df63057-30c5-4625-8365-9fe9907156c9","Type":"ContainerStarted","Data":"10fbf0e62dfde6687ba437dc0e79e6e5751716c471f39af601314d6e02d23f9f"} Dec 03 00:38:28 crc kubenswrapper[4953]: I1203 00:38:28.021625 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/prometheus-default-0" event={"ID":"603c47f8-e2e2-4dbd-aa98-a0eca1acf697","Type":"ContainerStarted","Data":"21b773aaaa4f653729d81e001ecc44715a5d49434db8e8be6dec9534d3452e15"} Dec 03 00:38:28 crc kubenswrapper[4953]: I1203 00:38:28.032240 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-coll-meter-smartgateway-7cd87f9766-bgbxc" event={"ID":"f2ba70b7-7676-47af-aa77-4f90579e7ec1","Type":"ContainerStarted","Data":"37a95652839919bde63d0e797655b4af5df4a89ce5e5658653dcd039d652b585"} Dec 03 00:38:28 crc kubenswrapper[4953]: I1203 00:38:28.035922 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-sens-meter-smartgateway-5759b4d97-nlgdm" event={"ID":"d660753e-48eb-4388-bf97-c7050b9cb1d2","Type":"ContainerStarted","Data":"b607e5fd2e405759332a7da7a924498a1b3c8971fa355ae76388d94df3114040"} Dec 03 00:38:28 crc kubenswrapper[4953]: I1203 00:38:28.908555 4953 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="service-telemetry/prometheus-default-0" Dec 03 00:38:29 crc kubenswrapper[4953]: I1203 00:38:29.055377 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-sens-meter-smartgateway-5759b4d97-nlgdm" event={"ID":"d660753e-48eb-4388-bf97-c7050b9cb1d2","Type":"ContainerStarted","Data":"aa68bf406674bc981bb5368ff89b464fd1d1b6951ed1cdc16c465b763db5b91c"} Dec 03 00:38:29 crc kubenswrapper[4953]: I1203 00:38:29.062087 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-coll-meter-smartgateway-7cd87f9766-bgbxc" event={"ID":"f2ba70b7-7676-47af-aa77-4f90579e7ec1","Type":"ContainerStarted","Data":"21d0a685363d735754b4a513c23061c91f2f9ab27c8eaeafc7daf0e0f8022506"} Dec 03 00:38:30 crc kubenswrapper[4953]: I1203 00:38:30.349049 4953 scope.go:117] "RemoveContainer" containerID="150ffc325f215c6da77fc72304c13cb10dafc800bbc932afd319cc34442a3f76" Dec 03 00:38:30 crc kubenswrapper[4953]: E1203 00:38:30.349699 4953 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6dfbn_openshift-machine-config-operator(53ffb1e9-9a2e-4869-80d7-776f4aa7ef8b)\"" pod="openshift-machine-config-operator/machine-config-daemon-6dfbn" podUID="53ffb1e9-9a2e-4869-80d7-776f4aa7ef8b" Dec 03 00:38:31 crc kubenswrapper[4953]: I1203 00:38:31.373787 4953 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="service-telemetry/prometheus-default-0" podStartSLOduration=7.279366232 podStartE2EDuration="45.373762041s" podCreationTimestamp="2025-12-03 00:37:46 +0000 UTC" firstStartedPulling="2025-12-03 00:37:49.148100994 +0000 UTC m=+1858.132381777" lastFinishedPulling="2025-12-03 00:38:27.242496803 +0000 UTC m=+1896.226777586" observedRunningTime="2025-12-03 00:38:28.062338547 +0000 UTC m=+1897.046619330" watchObservedRunningTime="2025-12-03 00:38:31.373762041 +0000 UTC m=+1900.358042824" Dec 03 00:38:33 crc kubenswrapper[4953]: I1203 00:38:33.905542 4953 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="service-telemetry/prometheus-default-0" Dec 03 00:38:33 crc kubenswrapper[4953]: I1203 00:38:33.942659 4953 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="service-telemetry/prometheus-default-0" Dec 03 00:38:34 crc kubenswrapper[4953]: I1203 00:38:34.149557 4953 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="service-telemetry/prometheus-default-0" Dec 03 00:38:37 crc kubenswrapper[4953]: I1203 00:38:37.956594 4953 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["service-telemetry/default-cloud1-coll-event-smartgateway-5bf48db5df-92twf"] Dec 03 00:38:37 crc kubenswrapper[4953]: I1203 00:38:37.959149 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/default-cloud1-coll-event-smartgateway-5bf48db5df-92twf" Dec 03 00:38:37 crc kubenswrapper[4953]: I1203 00:38:37.961875 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"default-cloud1-coll-event-sg-core-configmap" Dec 03 00:38:37 crc kubenswrapper[4953]: I1203 00:38:37.962509 4953 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"elasticsearch-es-cert" Dec 03 00:38:37 crc kubenswrapper[4953]: I1203 00:38:37.974899 4953 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/default-cloud1-coll-event-smartgateway-5bf48db5df-92twf"] Dec 03 00:38:38 crc kubenswrapper[4953]: I1203 00:38:38.046453 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v29h5\" (UniqueName: \"kubernetes.io/projected/9683141e-460a-4680-81ca-58907eb72789-kube-api-access-v29h5\") pod \"default-cloud1-coll-event-smartgateway-5bf48db5df-92twf\" (UID: \"9683141e-460a-4680-81ca-58907eb72789\") " pod="service-telemetry/default-cloud1-coll-event-smartgateway-5bf48db5df-92twf" Dec 03 00:38:38 crc kubenswrapper[4953]: I1203 00:38:38.046692 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-config\" (UniqueName: \"kubernetes.io/configmap/9683141e-460a-4680-81ca-58907eb72789-sg-core-config\") pod \"default-cloud1-coll-event-smartgateway-5bf48db5df-92twf\" (UID: \"9683141e-460a-4680-81ca-58907eb72789\") " pod="service-telemetry/default-cloud1-coll-event-smartgateway-5bf48db5df-92twf" Dec 03 00:38:38 crc kubenswrapper[4953]: I1203 00:38:38.046770 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"elastic-certs\" (UniqueName: \"kubernetes.io/secret/9683141e-460a-4680-81ca-58907eb72789-elastic-certs\") pod \"default-cloud1-coll-event-smartgateway-5bf48db5df-92twf\" (UID: \"9683141e-460a-4680-81ca-58907eb72789\") " pod="service-telemetry/default-cloud1-coll-event-smartgateway-5bf48db5df-92twf" Dec 03 00:38:38 crc kubenswrapper[4953]: I1203 00:38:38.046822 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/empty-dir/9683141e-460a-4680-81ca-58907eb72789-socket-dir\") pod \"default-cloud1-coll-event-smartgateway-5bf48db5df-92twf\" (UID: \"9683141e-460a-4680-81ca-58907eb72789\") " pod="service-telemetry/default-cloud1-coll-event-smartgateway-5bf48db5df-92twf" Dec 03 00:38:38 crc kubenswrapper[4953]: I1203 00:38:38.148209 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-config\" (UniqueName: \"kubernetes.io/configmap/9683141e-460a-4680-81ca-58907eb72789-sg-core-config\") pod \"default-cloud1-coll-event-smartgateway-5bf48db5df-92twf\" (UID: \"9683141e-460a-4680-81ca-58907eb72789\") " pod="service-telemetry/default-cloud1-coll-event-smartgateway-5bf48db5df-92twf" Dec 03 00:38:38 crc kubenswrapper[4953]: I1203 00:38:38.148304 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"elastic-certs\" (UniqueName: \"kubernetes.io/secret/9683141e-460a-4680-81ca-58907eb72789-elastic-certs\") pod \"default-cloud1-coll-event-smartgateway-5bf48db5df-92twf\" (UID: \"9683141e-460a-4680-81ca-58907eb72789\") " pod="service-telemetry/default-cloud1-coll-event-smartgateway-5bf48db5df-92twf" Dec 03 00:38:38 crc kubenswrapper[4953]: I1203 00:38:38.148339 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/empty-dir/9683141e-460a-4680-81ca-58907eb72789-socket-dir\") pod \"default-cloud1-coll-event-smartgateway-5bf48db5df-92twf\" (UID: \"9683141e-460a-4680-81ca-58907eb72789\") " pod="service-telemetry/default-cloud1-coll-event-smartgateway-5bf48db5df-92twf" Dec 03 00:38:38 crc kubenswrapper[4953]: I1203 00:38:38.148385 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v29h5\" (UniqueName: \"kubernetes.io/projected/9683141e-460a-4680-81ca-58907eb72789-kube-api-access-v29h5\") pod \"default-cloud1-coll-event-smartgateway-5bf48db5df-92twf\" (UID: \"9683141e-460a-4680-81ca-58907eb72789\") " pod="service-telemetry/default-cloud1-coll-event-smartgateway-5bf48db5df-92twf" Dec 03 00:38:38 crc kubenswrapper[4953]: I1203 00:38:38.149140 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/empty-dir/9683141e-460a-4680-81ca-58907eb72789-socket-dir\") pod \"default-cloud1-coll-event-smartgateway-5bf48db5df-92twf\" (UID: \"9683141e-460a-4680-81ca-58907eb72789\") " pod="service-telemetry/default-cloud1-coll-event-smartgateway-5bf48db5df-92twf" Dec 03 00:38:38 crc kubenswrapper[4953]: I1203 00:38:38.149393 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-config\" (UniqueName: \"kubernetes.io/configmap/9683141e-460a-4680-81ca-58907eb72789-sg-core-config\") pod \"default-cloud1-coll-event-smartgateway-5bf48db5df-92twf\" (UID: \"9683141e-460a-4680-81ca-58907eb72789\") " pod="service-telemetry/default-cloud1-coll-event-smartgateway-5bf48db5df-92twf" Dec 03 00:38:38 crc kubenswrapper[4953]: I1203 00:38:38.165678 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/alertmanager-default-0" event={"ID":"1df63057-30c5-4625-8365-9fe9907156c9","Type":"ContainerStarted","Data":"d8cf730124b15581d9a45a613d7e9e3ecdd5097625050f650bfcdbd9480ef20e"} Dec 03 00:38:38 crc kubenswrapper[4953]: I1203 00:38:38.167883 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"elastic-certs\" (UniqueName: \"kubernetes.io/secret/9683141e-460a-4680-81ca-58907eb72789-elastic-certs\") pod \"default-cloud1-coll-event-smartgateway-5bf48db5df-92twf\" (UID: \"9683141e-460a-4680-81ca-58907eb72789\") " pod="service-telemetry/default-cloud1-coll-event-smartgateway-5bf48db5df-92twf" Dec 03 00:38:38 crc kubenswrapper[4953]: I1203 00:38:38.169610 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v29h5\" (UniqueName: \"kubernetes.io/projected/9683141e-460a-4680-81ca-58907eb72789-kube-api-access-v29h5\") pod \"default-cloud1-coll-event-smartgateway-5bf48db5df-92twf\" (UID: \"9683141e-460a-4680-81ca-58907eb72789\") " pod="service-telemetry/default-cloud1-coll-event-smartgateway-5bf48db5df-92twf" Dec 03 00:38:38 crc kubenswrapper[4953]: I1203 00:38:38.668782 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/default-cloud1-coll-event-smartgateway-5bf48db5df-92twf" Dec 03 00:38:40 crc kubenswrapper[4953]: I1203 00:38:40.783304 4953 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["service-telemetry/default-cloud1-ceil-event-smartgateway-5967d7cb84-dgpqt"] Dec 03 00:38:40 crc kubenswrapper[4953]: I1203 00:38:40.785619 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/default-cloud1-ceil-event-smartgateway-5967d7cb84-dgpqt" Dec 03 00:38:40 crc kubenswrapper[4953]: I1203 00:38:40.789300 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"default-cloud1-ceil-event-sg-core-configmap" Dec 03 00:38:40 crc kubenswrapper[4953]: I1203 00:38:40.804276 4953 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/default-cloud1-ceil-event-smartgateway-5967d7cb84-dgpqt"] Dec 03 00:38:40 crc kubenswrapper[4953]: I1203 00:38:40.926884 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-config\" (UniqueName: \"kubernetes.io/configmap/b9fb7d41-3caf-41bc-ac33-23d6757b4b11-sg-core-config\") pod \"default-cloud1-ceil-event-smartgateway-5967d7cb84-dgpqt\" (UID: \"b9fb7d41-3caf-41bc-ac33-23d6757b4b11\") " pod="service-telemetry/default-cloud1-ceil-event-smartgateway-5967d7cb84-dgpqt" Dec 03 00:38:40 crc kubenswrapper[4953]: I1203 00:38:40.927004 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/empty-dir/b9fb7d41-3caf-41bc-ac33-23d6757b4b11-socket-dir\") pod \"default-cloud1-ceil-event-smartgateway-5967d7cb84-dgpqt\" (UID: \"b9fb7d41-3caf-41bc-ac33-23d6757b4b11\") " pod="service-telemetry/default-cloud1-ceil-event-smartgateway-5967d7cb84-dgpqt" Dec 03 00:38:40 crc kubenswrapper[4953]: I1203 00:38:40.927032 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rrbc8\" (UniqueName: \"kubernetes.io/projected/b9fb7d41-3caf-41bc-ac33-23d6757b4b11-kube-api-access-rrbc8\") pod \"default-cloud1-ceil-event-smartgateway-5967d7cb84-dgpqt\" (UID: \"b9fb7d41-3caf-41bc-ac33-23d6757b4b11\") " pod="service-telemetry/default-cloud1-ceil-event-smartgateway-5967d7cb84-dgpqt" Dec 03 00:38:40 crc kubenswrapper[4953]: I1203 00:38:40.927067 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"elastic-certs\" (UniqueName: \"kubernetes.io/secret/b9fb7d41-3caf-41bc-ac33-23d6757b4b11-elastic-certs\") pod \"default-cloud1-ceil-event-smartgateway-5967d7cb84-dgpqt\" (UID: \"b9fb7d41-3caf-41bc-ac33-23d6757b4b11\") " pod="service-telemetry/default-cloud1-ceil-event-smartgateway-5967d7cb84-dgpqt" Dec 03 00:38:41 crc kubenswrapper[4953]: I1203 00:38:41.028450 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-config\" (UniqueName: \"kubernetes.io/configmap/b9fb7d41-3caf-41bc-ac33-23d6757b4b11-sg-core-config\") pod \"default-cloud1-ceil-event-smartgateway-5967d7cb84-dgpqt\" (UID: \"b9fb7d41-3caf-41bc-ac33-23d6757b4b11\") " pod="service-telemetry/default-cloud1-ceil-event-smartgateway-5967d7cb84-dgpqt" Dec 03 00:38:41 crc kubenswrapper[4953]: I1203 00:38:41.028553 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/empty-dir/b9fb7d41-3caf-41bc-ac33-23d6757b4b11-socket-dir\") pod \"default-cloud1-ceil-event-smartgateway-5967d7cb84-dgpqt\" (UID: \"b9fb7d41-3caf-41bc-ac33-23d6757b4b11\") " pod="service-telemetry/default-cloud1-ceil-event-smartgateway-5967d7cb84-dgpqt" Dec 03 00:38:41 crc kubenswrapper[4953]: I1203 00:38:41.028574 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rrbc8\" (UniqueName: \"kubernetes.io/projected/b9fb7d41-3caf-41bc-ac33-23d6757b4b11-kube-api-access-rrbc8\") pod \"default-cloud1-ceil-event-smartgateway-5967d7cb84-dgpqt\" (UID: \"b9fb7d41-3caf-41bc-ac33-23d6757b4b11\") " pod="service-telemetry/default-cloud1-ceil-event-smartgateway-5967d7cb84-dgpqt" Dec 03 00:38:41 crc kubenswrapper[4953]: I1203 00:38:41.028601 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"elastic-certs\" (UniqueName: \"kubernetes.io/secret/b9fb7d41-3caf-41bc-ac33-23d6757b4b11-elastic-certs\") pod \"default-cloud1-ceil-event-smartgateway-5967d7cb84-dgpqt\" (UID: \"b9fb7d41-3caf-41bc-ac33-23d6757b4b11\") " pod="service-telemetry/default-cloud1-ceil-event-smartgateway-5967d7cb84-dgpqt" Dec 03 00:38:41 crc kubenswrapper[4953]: I1203 00:38:41.030007 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/empty-dir/b9fb7d41-3caf-41bc-ac33-23d6757b4b11-socket-dir\") pod \"default-cloud1-ceil-event-smartgateway-5967d7cb84-dgpqt\" (UID: \"b9fb7d41-3caf-41bc-ac33-23d6757b4b11\") " pod="service-telemetry/default-cloud1-ceil-event-smartgateway-5967d7cb84-dgpqt" Dec 03 00:38:41 crc kubenswrapper[4953]: I1203 00:38:41.030462 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-config\" (UniqueName: \"kubernetes.io/configmap/b9fb7d41-3caf-41bc-ac33-23d6757b4b11-sg-core-config\") pod \"default-cloud1-ceil-event-smartgateway-5967d7cb84-dgpqt\" (UID: \"b9fb7d41-3caf-41bc-ac33-23d6757b4b11\") " pod="service-telemetry/default-cloud1-ceil-event-smartgateway-5967d7cb84-dgpqt" Dec 03 00:38:41 crc kubenswrapper[4953]: I1203 00:38:41.040454 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"elastic-certs\" (UniqueName: \"kubernetes.io/secret/b9fb7d41-3caf-41bc-ac33-23d6757b4b11-elastic-certs\") pod \"default-cloud1-ceil-event-smartgateway-5967d7cb84-dgpqt\" (UID: \"b9fb7d41-3caf-41bc-ac33-23d6757b4b11\") " pod="service-telemetry/default-cloud1-ceil-event-smartgateway-5967d7cb84-dgpqt" Dec 03 00:38:41 crc kubenswrapper[4953]: I1203 00:38:41.060786 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rrbc8\" (UniqueName: \"kubernetes.io/projected/b9fb7d41-3caf-41bc-ac33-23d6757b4b11-kube-api-access-rrbc8\") pod \"default-cloud1-ceil-event-smartgateway-5967d7cb84-dgpqt\" (UID: \"b9fb7d41-3caf-41bc-ac33-23d6757b4b11\") " pod="service-telemetry/default-cloud1-ceil-event-smartgateway-5967d7cb84-dgpqt" Dec 03 00:38:41 crc kubenswrapper[4953]: I1203 00:38:41.117027 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/default-cloud1-ceil-event-smartgateway-5967d7cb84-dgpqt" Dec 03 00:38:43 crc kubenswrapper[4953]: I1203 00:38:43.553634 4953 scope.go:117] "RemoveContainer" containerID="150ffc325f215c6da77fc72304c13cb10dafc800bbc932afd319cc34442a3f76" Dec 03 00:38:43 crc kubenswrapper[4953]: E1203 00:38:43.554486 4953 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6dfbn_openshift-machine-config-operator(53ffb1e9-9a2e-4869-80d7-776f4aa7ef8b)\"" pod="openshift-machine-config-operator/machine-config-daemon-6dfbn" podUID="53ffb1e9-9a2e-4869-80d7-776f4aa7ef8b" Dec 03 00:38:43 crc kubenswrapper[4953]: E1203 00:38:43.810910 4953 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="image-registry.openshift-image-registry.svc:5000/service-telemetry/sg-bridge:latest" Dec 03 00:38:43 crc kubenswrapper[4953]: E1203 00:38:43.811176 4953 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:bridge,Image:image-registry.openshift-image-registry.svc:5000/service-telemetry/sg-bridge:latest,Command:[],Args:[--amqp_url amqp://default-interconnect.service-telemetry.svc.cluster.local:5673/collectd/cloud1-telemetry --block --stat_period 60 --rbc 15000 --count 0 --gw_unix /tmp/smartgateway --rbs 16384],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:socket-dir,ReadOnly:false,MountPath:/tmp,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-7k86r,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000670000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod default-cloud1-coll-meter-smartgateway-7cd87f9766-bgbxc_service-telemetry(f2ba70b7-7676-47af-aa77-4f90579e7ec1): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 03 00:38:43 crc kubenswrapper[4953]: E1203 00:38:43.818795 4953 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="image-registry.openshift-image-registry.svc:5000/service-telemetry/sg-bridge:latest" Dec 03 00:38:43 crc kubenswrapper[4953]: E1203 00:38:43.819018 4953 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:bridge,Image:image-registry.openshift-image-registry.svc:5000/service-telemetry/sg-bridge:latest,Command:[],Args:[--amqp_url amqp://default-interconnect.service-telemetry.svc.cluster.local:5673/anycast/ceilometer/cloud1-metering.sample --block --stat_period 60 --rbc 15000 --count 0 --gw_unix /tmp/smartgateway --rbs 16384],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:socket-dir,ReadOnly:false,MountPath:/tmp,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-4xn4p,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000670000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod default-cloud1-ceil-meter-smartgateway-57948895dc-blgcn_service-telemetry(43d17ba9-6fdf-4dca-991c-924f350daa0b): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 03 00:38:43 crc kubenswrapper[4953]: E1203 00:38:43.841369 4953 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="image-registry.openshift-image-registry.svc:5000/service-telemetry/sg-bridge:latest" Dec 03 00:38:43 crc kubenswrapper[4953]: E1203 00:38:43.841525 4953 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:bridge,Image:image-registry.openshift-image-registry.svc:5000/service-telemetry/sg-bridge:latest,Command:[],Args:[--amqp_url amqp://default-interconnect.service-telemetry.svc.cluster.local:5673/sensubility/cloud1-telemetry --block --stat_period 60 --rbc 15000 --count 0 --gw_unix /tmp/smartgateway --rbs 65535],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:socket-dir,ReadOnly:false,MountPath:/tmp,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-qclmr,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000670000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod default-cloud1-sens-meter-smartgateway-5759b4d97-nlgdm_service-telemetry(d660753e-48eb-4388-bf97-c7050b9cb1d2): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 03 00:38:44 crc kubenswrapper[4953]: I1203 00:38:44.356881 4953 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/default-cloud1-ceil-event-smartgateway-5967d7cb84-dgpqt"] Dec 03 00:38:44 crc kubenswrapper[4953]: W1203 00:38:44.366783 4953 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb9fb7d41_3caf_41bc_ac33_23d6757b4b11.slice/crio-48f18cb4381dcfc026b181b676bc3641c7cdf677275cfc382359cc8d3ca4ac3c WatchSource:0}: Error finding container 48f18cb4381dcfc026b181b676bc3641c7cdf677275cfc382359cc8d3ca4ac3c: Status 404 returned error can't find the container with id 48f18cb4381dcfc026b181b676bc3641c7cdf677275cfc382359cc8d3ca4ac3c Dec 03 00:38:44 crc kubenswrapper[4953]: I1203 00:38:44.480438 4953 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/default-cloud1-coll-event-smartgateway-5bf48db5df-92twf"] Dec 03 00:38:44 crc kubenswrapper[4953]: I1203 00:38:44.569154 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/alertmanager-default-0" event={"ID":"1df63057-30c5-4625-8365-9fe9907156c9","Type":"ContainerStarted","Data":"04490435b64d4669bac96e3dae7d221830ef869519ef0a4e378cd03d82279fd9"} Dec 03 00:38:44 crc kubenswrapper[4953]: I1203 00:38:44.570868 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-ceil-event-smartgateway-5967d7cb84-dgpqt" event={"ID":"b9fb7d41-3caf-41bc-ac33-23d6757b4b11","Type":"ContainerStarted","Data":"48f18cb4381dcfc026b181b676bc3641c7cdf677275cfc382359cc8d3ca4ac3c"} Dec 03 00:38:44 crc kubenswrapper[4953]: I1203 00:38:44.574053 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-coll-event-smartgateway-5bf48db5df-92twf" event={"ID":"9683141e-460a-4680-81ca-58907eb72789","Type":"ContainerStarted","Data":"a55be335bf226e7b04432459e5e3b0d467baf7acc0a8d1ac06b2c0b8d500a6c6"} Dec 03 00:38:44 crc kubenswrapper[4953]: I1203 00:38:44.598804 4953 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="service-telemetry/alertmanager-default-0" podStartSLOduration=20.657086661 podStartE2EDuration="44.598771366s" podCreationTimestamp="2025-12-03 00:38:00 +0000 UTC" firstStartedPulling="2025-12-03 00:38:19.920932929 +0000 UTC m=+1888.905213712" lastFinishedPulling="2025-12-03 00:38:43.862617634 +0000 UTC m=+1912.846898417" observedRunningTime="2025-12-03 00:38:44.596330275 +0000 UTC m=+1913.580611058" watchObservedRunningTime="2025-12-03 00:38:44.598771366 +0000 UTC m=+1913.583052149" Dec 03 00:38:45 crc kubenswrapper[4953]: I1203 00:38:45.604286 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-ceil-event-smartgateway-5967d7cb84-dgpqt" event={"ID":"b9fb7d41-3caf-41bc-ac33-23d6757b4b11","Type":"ContainerStarted","Data":"6a232af851bb739cff6a56d4c0f04ca57a10598fa96651f50704f0b8d17dabd1"} Dec 03 00:38:45 crc kubenswrapper[4953]: I1203 00:38:45.609769 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-coll-event-smartgateway-5bf48db5df-92twf" event={"ID":"9683141e-460a-4680-81ca-58907eb72789","Type":"ContainerStarted","Data":"5a4529673a6bfd10d36614b0f2c5f3fdcdfc9df36b4a246d41f4c3fa00d13db8"} Dec 03 00:38:56 crc kubenswrapper[4953]: E1203 00:38:56.602350 4953 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"bridge\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-57948895dc-blgcn" podUID="43d17ba9-6fdf-4dca-991c-924f350daa0b" Dec 03 00:38:56 crc kubenswrapper[4953]: E1203 00:38:56.603069 4953 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"bridge\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="service-telemetry/default-cloud1-sens-meter-smartgateway-5759b4d97-nlgdm" podUID="d660753e-48eb-4388-bf97-c7050b9cb1d2" Dec 03 00:38:56 crc kubenswrapper[4953]: E1203 00:38:56.608865 4953 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"bridge\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="service-telemetry/default-cloud1-coll-meter-smartgateway-7cd87f9766-bgbxc" podUID="f2ba70b7-7676-47af-aa77-4f90579e7ec1" Dec 03 00:38:56 crc kubenswrapper[4953]: I1203 00:38:56.976409 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-57948895dc-blgcn" event={"ID":"43d17ba9-6fdf-4dca-991c-924f350daa0b","Type":"ContainerStarted","Data":"697347ed6dde7b27f4854d26840288deb82f28ebf5257ff8ae6fb2ecd84b9754"} Dec 03 00:38:56 crc kubenswrapper[4953]: I1203 00:38:56.979144 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-coll-meter-smartgateway-7cd87f9766-bgbxc" event={"ID":"f2ba70b7-7676-47af-aa77-4f90579e7ec1","Type":"ContainerStarted","Data":"e5a516ee2ca524663d7125ac6c80eab2bf92b307bda12ad86715bd8ccd8aa1ad"} Dec 03 00:38:56 crc kubenswrapper[4953]: I1203 00:38:56.982463 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-sens-meter-smartgateway-5759b4d97-nlgdm" event={"ID":"d660753e-48eb-4388-bf97-c7050b9cb1d2","Type":"ContainerStarted","Data":"523a87a16aa47c54f859ce4e78f743cdab10315f9eabe5852965ed0b3b1bb508"} Dec 03 00:38:56 crc kubenswrapper[4953]: I1203 00:38:56.991016 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-ceil-event-smartgateway-5967d7cb84-dgpqt" event={"ID":"b9fb7d41-3caf-41bc-ac33-23d6757b4b11","Type":"ContainerStarted","Data":"6e72ceff85e5a8b148b2c56eaa31d079eb10b4599c5fb64fd98d768d255d0f7e"} Dec 03 00:38:57 crc kubenswrapper[4953]: I1203 00:38:57.000086 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-coll-event-smartgateway-5bf48db5df-92twf" event={"ID":"9683141e-460a-4680-81ca-58907eb72789","Type":"ContainerStarted","Data":"477a32b75b10803e1c690eb9f5a1f3a8933a5cd53e5e2d659d10d7322e42e706"} Dec 03 00:38:57 crc kubenswrapper[4953]: I1203 00:38:57.024991 4953 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="service-telemetry/default-cloud1-coll-event-smartgateway-5bf48db5df-92twf" podStartSLOduration=8.281518976 podStartE2EDuration="20.02494127s" podCreationTimestamp="2025-12-03 00:38:37 +0000 UTC" firstStartedPulling="2025-12-03 00:38:44.498939895 +0000 UTC m=+1913.483220678" lastFinishedPulling="2025-12-03 00:38:56.242362199 +0000 UTC m=+1925.226642972" observedRunningTime="2025-12-03 00:38:57.023490314 +0000 UTC m=+1926.007771097" watchObservedRunningTime="2025-12-03 00:38:57.02494127 +0000 UTC m=+1926.009222053" Dec 03 00:38:57 crc kubenswrapper[4953]: I1203 00:38:57.092387 4953 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="service-telemetry/default-cloud1-ceil-event-smartgateway-5967d7cb84-dgpqt" podStartSLOduration=5.223884599 podStartE2EDuration="17.092367359s" podCreationTimestamp="2025-12-03 00:38:40 +0000 UTC" firstStartedPulling="2025-12-03 00:38:44.374054283 +0000 UTC m=+1913.358335066" lastFinishedPulling="2025-12-03 00:38:56.242537043 +0000 UTC m=+1925.226817826" observedRunningTime="2025-12-03 00:38:57.086265978 +0000 UTC m=+1926.070546761" watchObservedRunningTime="2025-12-03 00:38:57.092367359 +0000 UTC m=+1926.076648142" Dec 03 00:38:57 crc kubenswrapper[4953]: I1203 00:38:57.348846 4953 scope.go:117] "RemoveContainer" containerID="150ffc325f215c6da77fc72304c13cb10dafc800bbc932afd319cc34442a3f76" Dec 03 00:38:57 crc kubenswrapper[4953]: E1203 00:38:57.349202 4953 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6dfbn_openshift-machine-config-operator(53ffb1e9-9a2e-4869-80d7-776f4aa7ef8b)\"" pod="openshift-machine-config-operator/machine-config-daemon-6dfbn" podUID="53ffb1e9-9a2e-4869-80d7-776f4aa7ef8b" Dec 03 00:38:58 crc kubenswrapper[4953]: I1203 00:38:58.010512 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-coll-meter-smartgateway-7cd87f9766-bgbxc" event={"ID":"f2ba70b7-7676-47af-aa77-4f90579e7ec1","Type":"ContainerStarted","Data":"4c8bffbb5e5d466c34012f2a1be704ab13e99c3adacdcd7da002338805fe6103"} Dec 03 00:38:58 crc kubenswrapper[4953]: I1203 00:38:58.016385 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-sens-meter-smartgateway-5759b4d97-nlgdm" event={"ID":"d660753e-48eb-4388-bf97-c7050b9cb1d2","Type":"ContainerStarted","Data":"f1d5f0a17b952cfeea649d1a3720f46609c494b24897944c298cfd5bc6dc36fb"} Dec 03 00:38:58 crc kubenswrapper[4953]: I1203 00:38:58.020366 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-57948895dc-blgcn" event={"ID":"43d17ba9-6fdf-4dca-991c-924f350daa0b","Type":"ContainerStarted","Data":"2bb8b24462f1cfa69afca915f9ace8fe56a68cf376237b167ebdbd8f0cf0a499"} Dec 03 00:38:58 crc kubenswrapper[4953]: I1203 00:38:58.047170 4953 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="service-telemetry/default-cloud1-coll-meter-smartgateway-7cd87f9766-bgbxc" podStartSLOduration=10.145616837 podStartE2EDuration="40.047140781s" podCreationTimestamp="2025-12-03 00:38:18 +0000 UTC" firstStartedPulling="2025-12-03 00:38:27.250066501 +0000 UTC m=+1896.234347284" lastFinishedPulling="2025-12-03 00:38:57.151590445 +0000 UTC m=+1926.135871228" observedRunningTime="2025-12-03 00:38:58.042609229 +0000 UTC m=+1927.026890012" watchObservedRunningTime="2025-12-03 00:38:58.047140781 +0000 UTC m=+1927.031421564" Dec 03 00:38:58 crc kubenswrapper[4953]: I1203 00:38:58.084144 4953 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-57948895dc-blgcn" podStartSLOduration=5.512021992 podStartE2EDuration="37.084115796s" podCreationTimestamp="2025-12-03 00:38:21 +0000 UTC" firstStartedPulling="2025-12-03 00:38:25.59725311 +0000 UTC m=+1894.581533893" lastFinishedPulling="2025-12-03 00:38:57.169346914 +0000 UTC m=+1926.153627697" observedRunningTime="2025-12-03 00:38:58.069836913 +0000 UTC m=+1927.054117696" watchObservedRunningTime="2025-12-03 00:38:58.084115796 +0000 UTC m=+1927.068396579" Dec 03 00:38:58 crc kubenswrapper[4953]: I1203 00:38:58.103781 4953 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="service-telemetry/default-cloud1-sens-meter-smartgateway-5759b4d97-nlgdm" podStartSLOduration=3.825328371 podStartE2EDuration="33.103752612s" podCreationTimestamp="2025-12-03 00:38:25 +0000 UTC" firstStartedPulling="2025-12-03 00:38:27.875688986 +0000 UTC m=+1896.859969769" lastFinishedPulling="2025-12-03 00:38:57.154113227 +0000 UTC m=+1926.138394010" observedRunningTime="2025-12-03 00:38:58.099985739 +0000 UTC m=+1927.084266522" watchObservedRunningTime="2025-12-03 00:38:58.103752612 +0000 UTC m=+1927.088033395" Dec 03 00:38:59 crc kubenswrapper[4953]: I1203 00:38:59.361179 4953 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["service-telemetry/default-interconnect-68864d46cb-mv5rb"] Dec 03 00:38:59 crc kubenswrapper[4953]: I1203 00:38:59.361483 4953 kuberuntime_container.go:808] "Killing container with a grace period" pod="service-telemetry/default-interconnect-68864d46cb-mv5rb" podUID="7e7fa27e-b88d-4d1c-b493-63b8f811dff5" containerName="default-interconnect" containerID="cri-o://e5a15270b4aaa460870081d673e9a5c7305c0ac209e59b6e719624ffb24b521f" gracePeriod=30 Dec 03 00:38:59 crc kubenswrapper[4953]: I1203 00:38:59.876934 4953 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/default-interconnect-68864d46cb-mv5rb" Dec 03 00:39:00 crc kubenswrapper[4953]: I1203 00:39:00.009000 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sasl-config\" (UniqueName: \"kubernetes.io/configmap/7e7fa27e-b88d-4d1c-b493-63b8f811dff5-sasl-config\") pod \"7e7fa27e-b88d-4d1c-b493-63b8f811dff5\" (UID: \"7e7fa27e-b88d-4d1c-b493-63b8f811dff5\") " Dec 03 00:39:00 crc kubenswrapper[4953]: I1203 00:39:00.009080 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-shh7v\" (UniqueName: \"kubernetes.io/projected/7e7fa27e-b88d-4d1c-b493-63b8f811dff5-kube-api-access-shh7v\") pod \"7e7fa27e-b88d-4d1c-b493-63b8f811dff5\" (UID: \"7e7fa27e-b88d-4d1c-b493-63b8f811dff5\") " Dec 03 00:39:00 crc kubenswrapper[4953]: I1203 00:39:00.009106 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"default-interconnect-inter-router-credentials\" (UniqueName: \"kubernetes.io/secret/7e7fa27e-b88d-4d1c-b493-63b8f811dff5-default-interconnect-inter-router-credentials\") pod \"7e7fa27e-b88d-4d1c-b493-63b8f811dff5\" (UID: \"7e7fa27e-b88d-4d1c-b493-63b8f811dff5\") " Dec 03 00:39:00 crc kubenswrapper[4953]: I1203 00:39:00.009193 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"default-interconnect-openstack-credentials\" (UniqueName: \"kubernetes.io/secret/7e7fa27e-b88d-4d1c-b493-63b8f811dff5-default-interconnect-openstack-credentials\") pod \"7e7fa27e-b88d-4d1c-b493-63b8f811dff5\" (UID: \"7e7fa27e-b88d-4d1c-b493-63b8f811dff5\") " Dec 03 00:39:00 crc kubenswrapper[4953]: I1203 00:39:00.009343 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"default-interconnect-inter-router-ca\" (UniqueName: \"kubernetes.io/secret/7e7fa27e-b88d-4d1c-b493-63b8f811dff5-default-interconnect-inter-router-ca\") pod \"7e7fa27e-b88d-4d1c-b493-63b8f811dff5\" (UID: \"7e7fa27e-b88d-4d1c-b493-63b8f811dff5\") " Dec 03 00:39:00 crc kubenswrapper[4953]: I1203 00:39:00.009374 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sasl-users\" (UniqueName: \"kubernetes.io/secret/7e7fa27e-b88d-4d1c-b493-63b8f811dff5-sasl-users\") pod \"7e7fa27e-b88d-4d1c-b493-63b8f811dff5\" (UID: \"7e7fa27e-b88d-4d1c-b493-63b8f811dff5\") " Dec 03 00:39:00 crc kubenswrapper[4953]: I1203 00:39:00.009424 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"default-interconnect-openstack-ca\" (UniqueName: \"kubernetes.io/secret/7e7fa27e-b88d-4d1c-b493-63b8f811dff5-default-interconnect-openstack-ca\") pod \"7e7fa27e-b88d-4d1c-b493-63b8f811dff5\" (UID: \"7e7fa27e-b88d-4d1c-b493-63b8f811dff5\") " Dec 03 00:39:00 crc kubenswrapper[4953]: I1203 00:39:00.010072 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7e7fa27e-b88d-4d1c-b493-63b8f811dff5-sasl-config" (OuterVolumeSpecName: "sasl-config") pod "7e7fa27e-b88d-4d1c-b493-63b8f811dff5" (UID: "7e7fa27e-b88d-4d1c-b493-63b8f811dff5"). InnerVolumeSpecName "sasl-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:39:00 crc kubenswrapper[4953]: I1203 00:39:00.019741 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7e7fa27e-b88d-4d1c-b493-63b8f811dff5-sasl-users" (OuterVolumeSpecName: "sasl-users") pod "7e7fa27e-b88d-4d1c-b493-63b8f811dff5" (UID: "7e7fa27e-b88d-4d1c-b493-63b8f811dff5"). InnerVolumeSpecName "sasl-users". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:39:00 crc kubenswrapper[4953]: I1203 00:39:00.020773 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7e7fa27e-b88d-4d1c-b493-63b8f811dff5-kube-api-access-shh7v" (OuterVolumeSpecName: "kube-api-access-shh7v") pod "7e7fa27e-b88d-4d1c-b493-63b8f811dff5" (UID: "7e7fa27e-b88d-4d1c-b493-63b8f811dff5"). InnerVolumeSpecName "kube-api-access-shh7v". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:39:00 crc kubenswrapper[4953]: I1203 00:39:00.020960 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7e7fa27e-b88d-4d1c-b493-63b8f811dff5-default-interconnect-openstack-ca" (OuterVolumeSpecName: "default-interconnect-openstack-ca") pod "7e7fa27e-b88d-4d1c-b493-63b8f811dff5" (UID: "7e7fa27e-b88d-4d1c-b493-63b8f811dff5"). InnerVolumeSpecName "default-interconnect-openstack-ca". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:39:00 crc kubenswrapper[4953]: I1203 00:39:00.022072 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7e7fa27e-b88d-4d1c-b493-63b8f811dff5-default-interconnect-inter-router-ca" (OuterVolumeSpecName: "default-interconnect-inter-router-ca") pod "7e7fa27e-b88d-4d1c-b493-63b8f811dff5" (UID: "7e7fa27e-b88d-4d1c-b493-63b8f811dff5"). InnerVolumeSpecName "default-interconnect-inter-router-ca". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:39:00 crc kubenswrapper[4953]: I1203 00:39:00.024074 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7e7fa27e-b88d-4d1c-b493-63b8f811dff5-default-interconnect-openstack-credentials" (OuterVolumeSpecName: "default-interconnect-openstack-credentials") pod "7e7fa27e-b88d-4d1c-b493-63b8f811dff5" (UID: "7e7fa27e-b88d-4d1c-b493-63b8f811dff5"). InnerVolumeSpecName "default-interconnect-openstack-credentials". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:39:00 crc kubenswrapper[4953]: I1203 00:39:00.025239 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7e7fa27e-b88d-4d1c-b493-63b8f811dff5-default-interconnect-inter-router-credentials" (OuterVolumeSpecName: "default-interconnect-inter-router-credentials") pod "7e7fa27e-b88d-4d1c-b493-63b8f811dff5" (UID: "7e7fa27e-b88d-4d1c-b493-63b8f811dff5"). InnerVolumeSpecName "default-interconnect-inter-router-credentials". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:39:00 crc kubenswrapper[4953]: I1203 00:39:00.055696 4953 generic.go:334] "Generic (PLEG): container finished" podID="43d17ba9-6fdf-4dca-991c-924f350daa0b" containerID="2bb8b24462f1cfa69afca915f9ace8fe56a68cf376237b167ebdbd8f0cf0a499" exitCode=0 Dec 03 00:39:00 crc kubenswrapper[4953]: I1203 00:39:00.055809 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-57948895dc-blgcn" event={"ID":"43d17ba9-6fdf-4dca-991c-924f350daa0b","Type":"ContainerDied","Data":"2bb8b24462f1cfa69afca915f9ace8fe56a68cf376237b167ebdbd8f0cf0a499"} Dec 03 00:39:00 crc kubenswrapper[4953]: I1203 00:39:00.056948 4953 scope.go:117] "RemoveContainer" containerID="2bb8b24462f1cfa69afca915f9ace8fe56a68cf376237b167ebdbd8f0cf0a499" Dec 03 00:39:00 crc kubenswrapper[4953]: I1203 00:39:00.061185 4953 generic.go:334] "Generic (PLEG): container finished" podID="7e7fa27e-b88d-4d1c-b493-63b8f811dff5" containerID="e5a15270b4aaa460870081d673e9a5c7305c0ac209e59b6e719624ffb24b521f" exitCode=0 Dec 03 00:39:00 crc kubenswrapper[4953]: I1203 00:39:00.061344 4953 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/default-interconnect-68864d46cb-mv5rb" Dec 03 00:39:00 crc kubenswrapper[4953]: I1203 00:39:00.061529 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-interconnect-68864d46cb-mv5rb" event={"ID":"7e7fa27e-b88d-4d1c-b493-63b8f811dff5","Type":"ContainerDied","Data":"e5a15270b4aaa460870081d673e9a5c7305c0ac209e59b6e719624ffb24b521f"} Dec 03 00:39:00 crc kubenswrapper[4953]: I1203 00:39:00.061584 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-interconnect-68864d46cb-mv5rb" event={"ID":"7e7fa27e-b88d-4d1c-b493-63b8f811dff5","Type":"ContainerDied","Data":"fb50727c588b8a1ca5306e94e7ef1c5437bfc069fe80d94177f2880f6b69023c"} Dec 03 00:39:00 crc kubenswrapper[4953]: I1203 00:39:00.061610 4953 scope.go:117] "RemoveContainer" containerID="e5a15270b4aaa460870081d673e9a5c7305c0ac209e59b6e719624ffb24b521f" Dec 03 00:39:00 crc kubenswrapper[4953]: I1203 00:39:00.086114 4953 generic.go:334] "Generic (PLEG): container finished" podID="f2ba70b7-7676-47af-aa77-4f90579e7ec1" containerID="4c8bffbb5e5d466c34012f2a1be704ab13e99c3adacdcd7da002338805fe6103" exitCode=0 Dec 03 00:39:00 crc kubenswrapper[4953]: I1203 00:39:00.086258 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-coll-meter-smartgateway-7cd87f9766-bgbxc" event={"ID":"f2ba70b7-7676-47af-aa77-4f90579e7ec1","Type":"ContainerDied","Data":"4c8bffbb5e5d466c34012f2a1be704ab13e99c3adacdcd7da002338805fe6103"} Dec 03 00:39:00 crc kubenswrapper[4953]: I1203 00:39:00.087079 4953 scope.go:117] "RemoveContainer" containerID="4c8bffbb5e5d466c34012f2a1be704ab13e99c3adacdcd7da002338805fe6103" Dec 03 00:39:00 crc kubenswrapper[4953]: I1203 00:39:00.098058 4953 generic.go:334] "Generic (PLEG): container finished" podID="b9fb7d41-3caf-41bc-ac33-23d6757b4b11" containerID="6a232af851bb739cff6a56d4c0f04ca57a10598fa96651f50704f0b8d17dabd1" exitCode=0 Dec 03 00:39:00 crc kubenswrapper[4953]: I1203 00:39:00.098200 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-ceil-event-smartgateway-5967d7cb84-dgpqt" event={"ID":"b9fb7d41-3caf-41bc-ac33-23d6757b4b11","Type":"ContainerDied","Data":"6a232af851bb739cff6a56d4c0f04ca57a10598fa96651f50704f0b8d17dabd1"} Dec 03 00:39:00 crc kubenswrapper[4953]: I1203 00:39:00.098919 4953 scope.go:117] "RemoveContainer" containerID="6a232af851bb739cff6a56d4c0f04ca57a10598fa96651f50704f0b8d17dabd1" Dec 03 00:39:00 crc kubenswrapper[4953]: I1203 00:39:00.113053 4953 reconciler_common.go:293] "Volume detached for volume \"default-interconnect-openstack-credentials\" (UniqueName: \"kubernetes.io/secret/7e7fa27e-b88d-4d1c-b493-63b8f811dff5-default-interconnect-openstack-credentials\") on node \"crc\" DevicePath \"\"" Dec 03 00:39:00 crc kubenswrapper[4953]: I1203 00:39:00.113096 4953 reconciler_common.go:293] "Volume detached for volume \"default-interconnect-inter-router-ca\" (UniqueName: \"kubernetes.io/secret/7e7fa27e-b88d-4d1c-b493-63b8f811dff5-default-interconnect-inter-router-ca\") on node \"crc\" DevicePath \"\"" Dec 03 00:39:00 crc kubenswrapper[4953]: I1203 00:39:00.113118 4953 reconciler_common.go:293] "Volume detached for volume \"sasl-users\" (UniqueName: \"kubernetes.io/secret/7e7fa27e-b88d-4d1c-b493-63b8f811dff5-sasl-users\") on node \"crc\" DevicePath \"\"" Dec 03 00:39:00 crc kubenswrapper[4953]: I1203 00:39:00.113144 4953 reconciler_common.go:293] "Volume detached for volume \"default-interconnect-openstack-ca\" (UniqueName: \"kubernetes.io/secret/7e7fa27e-b88d-4d1c-b493-63b8f811dff5-default-interconnect-openstack-ca\") on node \"crc\" DevicePath \"\"" Dec 03 00:39:00 crc kubenswrapper[4953]: I1203 00:39:00.113190 4953 reconciler_common.go:293] "Volume detached for volume \"sasl-config\" (UniqueName: \"kubernetes.io/configmap/7e7fa27e-b88d-4d1c-b493-63b8f811dff5-sasl-config\") on node \"crc\" DevicePath \"\"" Dec 03 00:39:00 crc kubenswrapper[4953]: I1203 00:39:00.113239 4953 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-shh7v\" (UniqueName: \"kubernetes.io/projected/7e7fa27e-b88d-4d1c-b493-63b8f811dff5-kube-api-access-shh7v\") on node \"crc\" DevicePath \"\"" Dec 03 00:39:00 crc kubenswrapper[4953]: I1203 00:39:00.113256 4953 reconciler_common.go:293] "Volume detached for volume \"default-interconnect-inter-router-credentials\" (UniqueName: \"kubernetes.io/secret/7e7fa27e-b88d-4d1c-b493-63b8f811dff5-default-interconnect-inter-router-credentials\") on node \"crc\" DevicePath \"\"" Dec 03 00:39:00 crc kubenswrapper[4953]: I1203 00:39:00.142642 4953 generic.go:334] "Generic (PLEG): container finished" podID="9683141e-460a-4680-81ca-58907eb72789" containerID="5a4529673a6bfd10d36614b0f2c5f3fdcdfc9df36b4a246d41f4c3fa00d13db8" exitCode=0 Dec 03 00:39:00 crc kubenswrapper[4953]: I1203 00:39:00.142754 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-coll-event-smartgateway-5bf48db5df-92twf" event={"ID":"9683141e-460a-4680-81ca-58907eb72789","Type":"ContainerDied","Data":"5a4529673a6bfd10d36614b0f2c5f3fdcdfc9df36b4a246d41f4c3fa00d13db8"} Dec 03 00:39:00 crc kubenswrapper[4953]: I1203 00:39:00.143497 4953 scope.go:117] "RemoveContainer" containerID="5a4529673a6bfd10d36614b0f2c5f3fdcdfc9df36b4a246d41f4c3fa00d13db8" Dec 03 00:39:00 crc kubenswrapper[4953]: I1203 00:39:00.168087 4953 scope.go:117] "RemoveContainer" containerID="e5a15270b4aaa460870081d673e9a5c7305c0ac209e59b6e719624ffb24b521f" Dec 03 00:39:00 crc kubenswrapper[4953]: E1203 00:39:00.187483 4953 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e5a15270b4aaa460870081d673e9a5c7305c0ac209e59b6e719624ffb24b521f\": container with ID starting with e5a15270b4aaa460870081d673e9a5c7305c0ac209e59b6e719624ffb24b521f not found: ID does not exist" containerID="e5a15270b4aaa460870081d673e9a5c7305c0ac209e59b6e719624ffb24b521f" Dec 03 00:39:00 crc kubenswrapper[4953]: I1203 00:39:00.187545 4953 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e5a15270b4aaa460870081d673e9a5c7305c0ac209e59b6e719624ffb24b521f"} err="failed to get container status \"e5a15270b4aaa460870081d673e9a5c7305c0ac209e59b6e719624ffb24b521f\": rpc error: code = NotFound desc = could not find container \"e5a15270b4aaa460870081d673e9a5c7305c0ac209e59b6e719624ffb24b521f\": container with ID starting with e5a15270b4aaa460870081d673e9a5c7305c0ac209e59b6e719624ffb24b521f not found: ID does not exist" Dec 03 00:39:00 crc kubenswrapper[4953]: I1203 00:39:00.384033 4953 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["service-telemetry/default-interconnect-68864d46cb-mv5rb"] Dec 03 00:39:00 crc kubenswrapper[4953]: I1203 00:39:00.420957 4953 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["service-telemetry/default-interconnect-68864d46cb-mv5rb"] Dec 03 00:39:00 crc kubenswrapper[4953]: I1203 00:39:00.960208 4953 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["service-telemetry/default-interconnect-68864d46cb-4644b"] Dec 03 00:39:00 crc kubenswrapper[4953]: E1203 00:39:00.960702 4953 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7e7fa27e-b88d-4d1c-b493-63b8f811dff5" containerName="default-interconnect" Dec 03 00:39:00 crc kubenswrapper[4953]: I1203 00:39:00.960733 4953 state_mem.go:107] "Deleted CPUSet assignment" podUID="7e7fa27e-b88d-4d1c-b493-63b8f811dff5" containerName="default-interconnect" Dec 03 00:39:00 crc kubenswrapper[4953]: I1203 00:39:00.960929 4953 memory_manager.go:354] "RemoveStaleState removing state" podUID="7e7fa27e-b88d-4d1c-b493-63b8f811dff5" containerName="default-interconnect" Dec 03 00:39:00 crc kubenswrapper[4953]: I1203 00:39:00.961762 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/default-interconnect-68864d46cb-4644b" Dec 03 00:39:00 crc kubenswrapper[4953]: I1203 00:39:00.964450 4953 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"default-interconnect-users" Dec 03 00:39:00 crc kubenswrapper[4953]: I1203 00:39:00.964654 4953 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"default-interconnect-openstack-credentials" Dec 03 00:39:00 crc kubenswrapper[4953]: I1203 00:39:00.964840 4953 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"default-interconnect-openstack-ca" Dec 03 00:39:00 crc kubenswrapper[4953]: I1203 00:39:00.965159 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"default-interconnect-sasl-config" Dec 03 00:39:00 crc kubenswrapper[4953]: I1203 00:39:00.965276 4953 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"default-interconnect-dockercfg-d5qbv" Dec 03 00:39:00 crc kubenswrapper[4953]: I1203 00:39:00.965805 4953 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"default-interconnect-inter-router-credentials" Dec 03 00:39:00 crc kubenswrapper[4953]: I1203 00:39:00.966119 4953 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"default-interconnect-inter-router-ca" Dec 03 00:39:00 crc kubenswrapper[4953]: I1203 00:39:00.979460 4953 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/default-interconnect-68864d46cb-4644b"] Dec 03 00:39:01 crc kubenswrapper[4953]: I1203 00:39:01.040202 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n96qf\" (UniqueName: \"kubernetes.io/projected/122f9ca6-e8e4-45ac-bfc4-0eec579dee72-kube-api-access-n96qf\") pod \"default-interconnect-68864d46cb-4644b\" (UID: \"122f9ca6-e8e4-45ac-bfc4-0eec579dee72\") " pod="service-telemetry/default-interconnect-68864d46cb-4644b" Dec 03 00:39:01 crc kubenswrapper[4953]: I1203 00:39:01.040306 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-interconnect-openstack-ca\" (UniqueName: \"kubernetes.io/secret/122f9ca6-e8e4-45ac-bfc4-0eec579dee72-default-interconnect-openstack-ca\") pod \"default-interconnect-68864d46cb-4644b\" (UID: \"122f9ca6-e8e4-45ac-bfc4-0eec579dee72\") " pod="service-telemetry/default-interconnect-68864d46cb-4644b" Dec 03 00:39:01 crc kubenswrapper[4953]: I1203 00:39:01.040375 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-interconnect-openstack-credentials\" (UniqueName: \"kubernetes.io/secret/122f9ca6-e8e4-45ac-bfc4-0eec579dee72-default-interconnect-openstack-credentials\") pod \"default-interconnect-68864d46cb-4644b\" (UID: \"122f9ca6-e8e4-45ac-bfc4-0eec579dee72\") " pod="service-telemetry/default-interconnect-68864d46cb-4644b" Dec 03 00:39:01 crc kubenswrapper[4953]: I1203 00:39:01.040397 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sasl-users\" (UniqueName: \"kubernetes.io/secret/122f9ca6-e8e4-45ac-bfc4-0eec579dee72-sasl-users\") pod \"default-interconnect-68864d46cb-4644b\" (UID: \"122f9ca6-e8e4-45ac-bfc4-0eec579dee72\") " pod="service-telemetry/default-interconnect-68864d46cb-4644b" Dec 03 00:39:01 crc kubenswrapper[4953]: I1203 00:39:01.040415 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-interconnect-inter-router-credentials\" (UniqueName: \"kubernetes.io/secret/122f9ca6-e8e4-45ac-bfc4-0eec579dee72-default-interconnect-inter-router-credentials\") pod \"default-interconnect-68864d46cb-4644b\" (UID: \"122f9ca6-e8e4-45ac-bfc4-0eec579dee72\") " pod="service-telemetry/default-interconnect-68864d46cb-4644b" Dec 03 00:39:01 crc kubenswrapper[4953]: I1203 00:39:01.040438 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-interconnect-inter-router-ca\" (UniqueName: \"kubernetes.io/secret/122f9ca6-e8e4-45ac-bfc4-0eec579dee72-default-interconnect-inter-router-ca\") pod \"default-interconnect-68864d46cb-4644b\" (UID: \"122f9ca6-e8e4-45ac-bfc4-0eec579dee72\") " pod="service-telemetry/default-interconnect-68864d46cb-4644b" Dec 03 00:39:01 crc kubenswrapper[4953]: I1203 00:39:01.040462 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sasl-config\" (UniqueName: \"kubernetes.io/configmap/122f9ca6-e8e4-45ac-bfc4-0eec579dee72-sasl-config\") pod \"default-interconnect-68864d46cb-4644b\" (UID: \"122f9ca6-e8e4-45ac-bfc4-0eec579dee72\") " pod="service-telemetry/default-interconnect-68864d46cb-4644b" Dec 03 00:39:01 crc kubenswrapper[4953]: I1203 00:39:01.142705 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-interconnect-openstack-ca\" (UniqueName: \"kubernetes.io/secret/122f9ca6-e8e4-45ac-bfc4-0eec579dee72-default-interconnect-openstack-ca\") pod \"default-interconnect-68864d46cb-4644b\" (UID: \"122f9ca6-e8e4-45ac-bfc4-0eec579dee72\") " pod="service-telemetry/default-interconnect-68864d46cb-4644b" Dec 03 00:39:01 crc kubenswrapper[4953]: I1203 00:39:01.143197 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-interconnect-openstack-credentials\" (UniqueName: \"kubernetes.io/secret/122f9ca6-e8e4-45ac-bfc4-0eec579dee72-default-interconnect-openstack-credentials\") pod \"default-interconnect-68864d46cb-4644b\" (UID: \"122f9ca6-e8e4-45ac-bfc4-0eec579dee72\") " pod="service-telemetry/default-interconnect-68864d46cb-4644b" Dec 03 00:39:01 crc kubenswrapper[4953]: I1203 00:39:01.143326 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sasl-users\" (UniqueName: \"kubernetes.io/secret/122f9ca6-e8e4-45ac-bfc4-0eec579dee72-sasl-users\") pod \"default-interconnect-68864d46cb-4644b\" (UID: \"122f9ca6-e8e4-45ac-bfc4-0eec579dee72\") " pod="service-telemetry/default-interconnect-68864d46cb-4644b" Dec 03 00:39:01 crc kubenswrapper[4953]: I1203 00:39:01.143426 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-interconnect-inter-router-credentials\" (UniqueName: \"kubernetes.io/secret/122f9ca6-e8e4-45ac-bfc4-0eec579dee72-default-interconnect-inter-router-credentials\") pod \"default-interconnect-68864d46cb-4644b\" (UID: \"122f9ca6-e8e4-45ac-bfc4-0eec579dee72\") " pod="service-telemetry/default-interconnect-68864d46cb-4644b" Dec 03 00:39:01 crc kubenswrapper[4953]: I1203 00:39:01.143512 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-interconnect-inter-router-ca\" (UniqueName: \"kubernetes.io/secret/122f9ca6-e8e4-45ac-bfc4-0eec579dee72-default-interconnect-inter-router-ca\") pod \"default-interconnect-68864d46cb-4644b\" (UID: \"122f9ca6-e8e4-45ac-bfc4-0eec579dee72\") " pod="service-telemetry/default-interconnect-68864d46cb-4644b" Dec 03 00:39:01 crc kubenswrapper[4953]: I1203 00:39:01.143602 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sasl-config\" (UniqueName: \"kubernetes.io/configmap/122f9ca6-e8e4-45ac-bfc4-0eec579dee72-sasl-config\") pod \"default-interconnect-68864d46cb-4644b\" (UID: \"122f9ca6-e8e4-45ac-bfc4-0eec579dee72\") " pod="service-telemetry/default-interconnect-68864d46cb-4644b" Dec 03 00:39:01 crc kubenswrapper[4953]: I1203 00:39:01.143713 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n96qf\" (UniqueName: \"kubernetes.io/projected/122f9ca6-e8e4-45ac-bfc4-0eec579dee72-kube-api-access-n96qf\") pod \"default-interconnect-68864d46cb-4644b\" (UID: \"122f9ca6-e8e4-45ac-bfc4-0eec579dee72\") " pod="service-telemetry/default-interconnect-68864d46cb-4644b" Dec 03 00:39:01 crc kubenswrapper[4953]: I1203 00:39:01.144733 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sasl-config\" (UniqueName: \"kubernetes.io/configmap/122f9ca6-e8e4-45ac-bfc4-0eec579dee72-sasl-config\") pod \"default-interconnect-68864d46cb-4644b\" (UID: \"122f9ca6-e8e4-45ac-bfc4-0eec579dee72\") " pod="service-telemetry/default-interconnect-68864d46cb-4644b" Dec 03 00:39:01 crc kubenswrapper[4953]: I1203 00:39:01.149192 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-interconnect-inter-router-ca\" (UniqueName: \"kubernetes.io/secret/122f9ca6-e8e4-45ac-bfc4-0eec579dee72-default-interconnect-inter-router-ca\") pod \"default-interconnect-68864d46cb-4644b\" (UID: \"122f9ca6-e8e4-45ac-bfc4-0eec579dee72\") " pod="service-telemetry/default-interconnect-68864d46cb-4644b" Dec 03 00:39:01 crc kubenswrapper[4953]: I1203 00:39:01.149694 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-interconnect-openstack-credentials\" (UniqueName: \"kubernetes.io/secret/122f9ca6-e8e4-45ac-bfc4-0eec579dee72-default-interconnect-openstack-credentials\") pod \"default-interconnect-68864d46cb-4644b\" (UID: \"122f9ca6-e8e4-45ac-bfc4-0eec579dee72\") " pod="service-telemetry/default-interconnect-68864d46cb-4644b" Dec 03 00:39:01 crc kubenswrapper[4953]: I1203 00:39:01.150838 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-interconnect-inter-router-credentials\" (UniqueName: \"kubernetes.io/secret/122f9ca6-e8e4-45ac-bfc4-0eec579dee72-default-interconnect-inter-router-credentials\") pod \"default-interconnect-68864d46cb-4644b\" (UID: \"122f9ca6-e8e4-45ac-bfc4-0eec579dee72\") " pod="service-telemetry/default-interconnect-68864d46cb-4644b" Dec 03 00:39:01 crc kubenswrapper[4953]: I1203 00:39:01.151330 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sasl-users\" (UniqueName: \"kubernetes.io/secret/122f9ca6-e8e4-45ac-bfc4-0eec579dee72-sasl-users\") pod \"default-interconnect-68864d46cb-4644b\" (UID: \"122f9ca6-e8e4-45ac-bfc4-0eec579dee72\") " pod="service-telemetry/default-interconnect-68864d46cb-4644b" Dec 03 00:39:01 crc kubenswrapper[4953]: I1203 00:39:01.152286 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-interconnect-openstack-ca\" (UniqueName: \"kubernetes.io/secret/122f9ca6-e8e4-45ac-bfc4-0eec579dee72-default-interconnect-openstack-ca\") pod \"default-interconnect-68864d46cb-4644b\" (UID: \"122f9ca6-e8e4-45ac-bfc4-0eec579dee72\") " pod="service-telemetry/default-interconnect-68864d46cb-4644b" Dec 03 00:39:01 crc kubenswrapper[4953]: I1203 00:39:01.157077 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-coll-event-smartgateway-5bf48db5df-92twf" event={"ID":"9683141e-460a-4680-81ca-58907eb72789","Type":"ContainerStarted","Data":"d7b8179b1510c185c7e7579faefd9679e1cc2a3085b4dcddaa45bb459a502caa"} Dec 03 00:39:01 crc kubenswrapper[4953]: I1203 00:39:01.162196 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-57948895dc-blgcn" event={"ID":"43d17ba9-6fdf-4dca-991c-924f350daa0b","Type":"ContainerStarted","Data":"22c8df2d89ef32be5e4546e57460efdeffc11511a7493e4df9aa89d10a537710"} Dec 03 00:39:01 crc kubenswrapper[4953]: I1203 00:39:01.174540 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-coll-meter-smartgateway-7cd87f9766-bgbxc" event={"ID":"f2ba70b7-7676-47af-aa77-4f90579e7ec1","Type":"ContainerStarted","Data":"64e62958118a8c50351495310c184e1c9a0003e66823530ee93a94a3a0bbf56f"} Dec 03 00:39:01 crc kubenswrapper[4953]: I1203 00:39:01.175525 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n96qf\" (UniqueName: \"kubernetes.io/projected/122f9ca6-e8e4-45ac-bfc4-0eec579dee72-kube-api-access-n96qf\") pod \"default-interconnect-68864d46cb-4644b\" (UID: \"122f9ca6-e8e4-45ac-bfc4-0eec579dee72\") " pod="service-telemetry/default-interconnect-68864d46cb-4644b" Dec 03 00:39:01 crc kubenswrapper[4953]: I1203 00:39:01.186580 4953 generic.go:334] "Generic (PLEG): container finished" podID="d660753e-48eb-4388-bf97-c7050b9cb1d2" containerID="f1d5f0a17b952cfeea649d1a3720f46609c494b24897944c298cfd5bc6dc36fb" exitCode=0 Dec 03 00:39:01 crc kubenswrapper[4953]: I1203 00:39:01.186694 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-sens-meter-smartgateway-5759b4d97-nlgdm" event={"ID":"d660753e-48eb-4388-bf97-c7050b9cb1d2","Type":"ContainerDied","Data":"f1d5f0a17b952cfeea649d1a3720f46609c494b24897944c298cfd5bc6dc36fb"} Dec 03 00:39:01 crc kubenswrapper[4953]: I1203 00:39:01.189345 4953 scope.go:117] "RemoveContainer" containerID="f1d5f0a17b952cfeea649d1a3720f46609c494b24897944c298cfd5bc6dc36fb" Dec 03 00:39:01 crc kubenswrapper[4953]: I1203 00:39:01.189632 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-ceil-event-smartgateway-5967d7cb84-dgpqt" event={"ID":"b9fb7d41-3caf-41bc-ac33-23d6757b4b11","Type":"ContainerStarted","Data":"8459c45dddae86d9f97b4b43679ae5e69556a32c8d91a72b2d96967724932e45"} Dec 03 00:39:01 crc kubenswrapper[4953]: I1203 00:39:01.323692 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/default-interconnect-68864d46cb-4644b" Dec 03 00:39:01 crc kubenswrapper[4953]: I1203 00:39:01.376837 4953 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7e7fa27e-b88d-4d1c-b493-63b8f811dff5" path="/var/lib/kubelet/pods/7e7fa27e-b88d-4d1c-b493-63b8f811dff5/volumes" Dec 03 00:39:02 crc kubenswrapper[4953]: I1203 00:39:02.014040 4953 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/default-interconnect-68864d46cb-4644b"] Dec 03 00:39:02 crc kubenswrapper[4953]: W1203 00:39:02.025948 4953 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod122f9ca6_e8e4_45ac_bfc4_0eec579dee72.slice/crio-df3d69c4e2a7b3f61141487e84dc448cffdb3eadce5469fde554cb4470219690 WatchSource:0}: Error finding container df3d69c4e2a7b3f61141487e84dc448cffdb3eadce5469fde554cb4470219690: Status 404 returned error can't find the container with id df3d69c4e2a7b3f61141487e84dc448cffdb3eadce5469fde554cb4470219690 Dec 03 00:39:02 crc kubenswrapper[4953]: I1203 00:39:02.206198 4953 generic.go:334] "Generic (PLEG): container finished" podID="f2ba70b7-7676-47af-aa77-4f90579e7ec1" containerID="64e62958118a8c50351495310c184e1c9a0003e66823530ee93a94a3a0bbf56f" exitCode=0 Dec 03 00:39:02 crc kubenswrapper[4953]: I1203 00:39:02.206302 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-coll-meter-smartgateway-7cd87f9766-bgbxc" event={"ID":"f2ba70b7-7676-47af-aa77-4f90579e7ec1","Type":"ContainerDied","Data":"64e62958118a8c50351495310c184e1c9a0003e66823530ee93a94a3a0bbf56f"} Dec 03 00:39:02 crc kubenswrapper[4953]: I1203 00:39:02.206373 4953 scope.go:117] "RemoveContainer" containerID="4c8bffbb5e5d466c34012f2a1be704ab13e99c3adacdcd7da002338805fe6103" Dec 03 00:39:02 crc kubenswrapper[4953]: I1203 00:39:02.207428 4953 scope.go:117] "RemoveContainer" containerID="64e62958118a8c50351495310c184e1c9a0003e66823530ee93a94a3a0bbf56f" Dec 03 00:39:02 crc kubenswrapper[4953]: E1203 00:39:02.207876 4953 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"bridge\" with CrashLoopBackOff: \"back-off 10s restarting failed container=bridge pod=default-cloud1-coll-meter-smartgateway-7cd87f9766-bgbxc_service-telemetry(f2ba70b7-7676-47af-aa77-4f90579e7ec1)\"" pod="service-telemetry/default-cloud1-coll-meter-smartgateway-7cd87f9766-bgbxc" podUID="f2ba70b7-7676-47af-aa77-4f90579e7ec1" Dec 03 00:39:02 crc kubenswrapper[4953]: I1203 00:39:02.222319 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-sens-meter-smartgateway-5759b4d97-nlgdm" event={"ID":"d660753e-48eb-4388-bf97-c7050b9cb1d2","Type":"ContainerStarted","Data":"eeb6f10a572352c31fe4dc1894037f7459c6ff2ea7d5601fe9cecb9084034f7d"} Dec 03 00:39:02 crc kubenswrapper[4953]: I1203 00:39:02.230666 4953 generic.go:334] "Generic (PLEG): container finished" podID="b9fb7d41-3caf-41bc-ac33-23d6757b4b11" containerID="8459c45dddae86d9f97b4b43679ae5e69556a32c8d91a72b2d96967724932e45" exitCode=0 Dec 03 00:39:02 crc kubenswrapper[4953]: I1203 00:39:02.230700 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-ceil-event-smartgateway-5967d7cb84-dgpqt" event={"ID":"b9fb7d41-3caf-41bc-ac33-23d6757b4b11","Type":"ContainerDied","Data":"8459c45dddae86d9f97b4b43679ae5e69556a32c8d91a72b2d96967724932e45"} Dec 03 00:39:02 crc kubenswrapper[4953]: I1203 00:39:02.231087 4953 scope.go:117] "RemoveContainer" containerID="8459c45dddae86d9f97b4b43679ae5e69556a32c8d91a72b2d96967724932e45" Dec 03 00:39:02 crc kubenswrapper[4953]: E1203 00:39:02.231279 4953 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"bridge\" with CrashLoopBackOff: \"back-off 10s restarting failed container=bridge pod=default-cloud1-ceil-event-smartgateway-5967d7cb84-dgpqt_service-telemetry(b9fb7d41-3caf-41bc-ac33-23d6757b4b11)\"" pod="service-telemetry/default-cloud1-ceil-event-smartgateway-5967d7cb84-dgpqt" podUID="b9fb7d41-3caf-41bc-ac33-23d6757b4b11" Dec 03 00:39:02 crc kubenswrapper[4953]: I1203 00:39:02.244791 4953 generic.go:334] "Generic (PLEG): container finished" podID="9683141e-460a-4680-81ca-58907eb72789" containerID="d7b8179b1510c185c7e7579faefd9679e1cc2a3085b4dcddaa45bb459a502caa" exitCode=0 Dec 03 00:39:02 crc kubenswrapper[4953]: I1203 00:39:02.244913 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-coll-event-smartgateway-5bf48db5df-92twf" event={"ID":"9683141e-460a-4680-81ca-58907eb72789","Type":"ContainerDied","Data":"d7b8179b1510c185c7e7579faefd9679e1cc2a3085b4dcddaa45bb459a502caa"} Dec 03 00:39:02 crc kubenswrapper[4953]: I1203 00:39:02.245518 4953 scope.go:117] "RemoveContainer" containerID="d7b8179b1510c185c7e7579faefd9679e1cc2a3085b4dcddaa45bb459a502caa" Dec 03 00:39:02 crc kubenswrapper[4953]: E1203 00:39:02.245736 4953 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"bridge\" with CrashLoopBackOff: \"back-off 10s restarting failed container=bridge pod=default-cloud1-coll-event-smartgateway-5bf48db5df-92twf_service-telemetry(9683141e-460a-4680-81ca-58907eb72789)\"" pod="service-telemetry/default-cloud1-coll-event-smartgateway-5bf48db5df-92twf" podUID="9683141e-460a-4680-81ca-58907eb72789" Dec 03 00:39:02 crc kubenswrapper[4953]: I1203 00:39:02.264814 4953 generic.go:334] "Generic (PLEG): container finished" podID="43d17ba9-6fdf-4dca-991c-924f350daa0b" containerID="22c8df2d89ef32be5e4546e57460efdeffc11511a7493e4df9aa89d10a537710" exitCode=0 Dec 03 00:39:02 crc kubenswrapper[4953]: I1203 00:39:02.264949 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-57948895dc-blgcn" event={"ID":"43d17ba9-6fdf-4dca-991c-924f350daa0b","Type":"ContainerDied","Data":"22c8df2d89ef32be5e4546e57460efdeffc11511a7493e4df9aa89d10a537710"} Dec 03 00:39:02 crc kubenswrapper[4953]: I1203 00:39:02.265813 4953 scope.go:117] "RemoveContainer" containerID="22c8df2d89ef32be5e4546e57460efdeffc11511a7493e4df9aa89d10a537710" Dec 03 00:39:02 crc kubenswrapper[4953]: E1203 00:39:02.266068 4953 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"bridge\" with CrashLoopBackOff: \"back-off 10s restarting failed container=bridge pod=default-cloud1-ceil-meter-smartgateway-57948895dc-blgcn_service-telemetry(43d17ba9-6fdf-4dca-991c-924f350daa0b)\"" pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-57948895dc-blgcn" podUID="43d17ba9-6fdf-4dca-991c-924f350daa0b" Dec 03 00:39:02 crc kubenswrapper[4953]: I1203 00:39:02.272399 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-interconnect-68864d46cb-4644b" event={"ID":"122f9ca6-e8e4-45ac-bfc4-0eec579dee72","Type":"ContainerStarted","Data":"22e0601469f15ec3b44a11f3c3bef2b9889884462250585c51662ab7c377ada8"} Dec 03 00:39:02 crc kubenswrapper[4953]: I1203 00:39:02.272451 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-interconnect-68864d46cb-4644b" event={"ID":"122f9ca6-e8e4-45ac-bfc4-0eec579dee72","Type":"ContainerStarted","Data":"df3d69c4e2a7b3f61141487e84dc448cffdb3eadce5469fde554cb4470219690"} Dec 03 00:39:02 crc kubenswrapper[4953]: I1203 00:39:02.276093 4953 scope.go:117] "RemoveContainer" containerID="6a232af851bb739cff6a56d4c0f04ca57a10598fa96651f50704f0b8d17dabd1" Dec 03 00:39:02 crc kubenswrapper[4953]: I1203 00:39:02.339063 4953 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="service-telemetry/default-interconnect-68864d46cb-4644b" podStartSLOduration=3.339022224 podStartE2EDuration="3.339022224s" podCreationTimestamp="2025-12-03 00:38:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 00:39:02.327321744 +0000 UTC m=+1931.311602527" watchObservedRunningTime="2025-12-03 00:39:02.339022224 +0000 UTC m=+1931.323303017" Dec 03 00:39:02 crc kubenswrapper[4953]: I1203 00:39:02.348216 4953 scope.go:117] "RemoveContainer" containerID="5a4529673a6bfd10d36614b0f2c5f3fdcdfc9df36b4a246d41f4c3fa00d13db8" Dec 03 00:39:02 crc kubenswrapper[4953]: I1203 00:39:02.417893 4953 scope.go:117] "RemoveContainer" containerID="2bb8b24462f1cfa69afca915f9ace8fe56a68cf376237b167ebdbd8f0cf0a499" Dec 03 00:39:03 crc kubenswrapper[4953]: I1203 00:39:03.295612 4953 generic.go:334] "Generic (PLEG): container finished" podID="d660753e-48eb-4388-bf97-c7050b9cb1d2" containerID="eeb6f10a572352c31fe4dc1894037f7459c6ff2ea7d5601fe9cecb9084034f7d" exitCode=0 Dec 03 00:39:03 crc kubenswrapper[4953]: I1203 00:39:03.295815 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-sens-meter-smartgateway-5759b4d97-nlgdm" event={"ID":"d660753e-48eb-4388-bf97-c7050b9cb1d2","Type":"ContainerDied","Data":"eeb6f10a572352c31fe4dc1894037f7459c6ff2ea7d5601fe9cecb9084034f7d"} Dec 03 00:39:03 crc kubenswrapper[4953]: I1203 00:39:03.296340 4953 scope.go:117] "RemoveContainer" containerID="f1d5f0a17b952cfeea649d1a3720f46609c494b24897944c298cfd5bc6dc36fb" Dec 03 00:39:03 crc kubenswrapper[4953]: I1203 00:39:03.297141 4953 scope.go:117] "RemoveContainer" containerID="eeb6f10a572352c31fe4dc1894037f7459c6ff2ea7d5601fe9cecb9084034f7d" Dec 03 00:39:03 crc kubenswrapper[4953]: E1203 00:39:03.297415 4953 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"bridge\" with CrashLoopBackOff: \"back-off 10s restarting failed container=bridge pod=default-cloud1-sens-meter-smartgateway-5759b4d97-nlgdm_service-telemetry(d660753e-48eb-4388-bf97-c7050b9cb1d2)\"" pod="service-telemetry/default-cloud1-sens-meter-smartgateway-5759b4d97-nlgdm" podUID="d660753e-48eb-4388-bf97-c7050b9cb1d2" Dec 03 00:39:08 crc kubenswrapper[4953]: I1203 00:39:08.027566 4953 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["service-telemetry/qdr-test"] Dec 03 00:39:08 crc kubenswrapper[4953]: I1203 00:39:08.032824 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/qdr-test" Dec 03 00:39:08 crc kubenswrapper[4953]: I1203 00:39:08.045289 4953 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/qdr-test"] Dec 03 00:39:08 crc kubenswrapper[4953]: I1203 00:39:08.050406 4953 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"default-interconnect-selfsigned" Dec 03 00:39:08 crc kubenswrapper[4953]: I1203 00:39:08.051107 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"qdr-test-config" Dec 03 00:39:08 crc kubenswrapper[4953]: I1203 00:39:08.096197 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-interconnect-selfsigned-cert\" (UniqueName: \"kubernetes.io/secret/7298274a-5d57-47f7-b2a7-49869fc2f21c-default-interconnect-selfsigned-cert\") pod \"qdr-test\" (UID: \"7298274a-5d57-47f7-b2a7-49869fc2f21c\") " pod="service-telemetry/qdr-test" Dec 03 00:39:08 crc kubenswrapper[4953]: I1203 00:39:08.096266 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vqcb2\" (UniqueName: \"kubernetes.io/projected/7298274a-5d57-47f7-b2a7-49869fc2f21c-kube-api-access-vqcb2\") pod \"qdr-test\" (UID: \"7298274a-5d57-47f7-b2a7-49869fc2f21c\") " pod="service-telemetry/qdr-test" Dec 03 00:39:08 crc kubenswrapper[4953]: I1203 00:39:08.096303 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"qdr-test-config\" (UniqueName: \"kubernetes.io/configmap/7298274a-5d57-47f7-b2a7-49869fc2f21c-qdr-test-config\") pod \"qdr-test\" (UID: \"7298274a-5d57-47f7-b2a7-49869fc2f21c\") " pod="service-telemetry/qdr-test" Dec 03 00:39:08 crc kubenswrapper[4953]: I1203 00:39:08.198279 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-interconnect-selfsigned-cert\" (UniqueName: \"kubernetes.io/secret/7298274a-5d57-47f7-b2a7-49869fc2f21c-default-interconnect-selfsigned-cert\") pod \"qdr-test\" (UID: \"7298274a-5d57-47f7-b2a7-49869fc2f21c\") " pod="service-telemetry/qdr-test" Dec 03 00:39:08 crc kubenswrapper[4953]: I1203 00:39:08.198360 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vqcb2\" (UniqueName: \"kubernetes.io/projected/7298274a-5d57-47f7-b2a7-49869fc2f21c-kube-api-access-vqcb2\") pod \"qdr-test\" (UID: \"7298274a-5d57-47f7-b2a7-49869fc2f21c\") " pod="service-telemetry/qdr-test" Dec 03 00:39:08 crc kubenswrapper[4953]: I1203 00:39:08.198400 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"qdr-test-config\" (UniqueName: \"kubernetes.io/configmap/7298274a-5d57-47f7-b2a7-49869fc2f21c-qdr-test-config\") pod \"qdr-test\" (UID: \"7298274a-5d57-47f7-b2a7-49869fc2f21c\") " pod="service-telemetry/qdr-test" Dec 03 00:39:08 crc kubenswrapper[4953]: I1203 00:39:08.199988 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"qdr-test-config\" (UniqueName: \"kubernetes.io/configmap/7298274a-5d57-47f7-b2a7-49869fc2f21c-qdr-test-config\") pod \"qdr-test\" (UID: \"7298274a-5d57-47f7-b2a7-49869fc2f21c\") " pod="service-telemetry/qdr-test" Dec 03 00:39:08 crc kubenswrapper[4953]: I1203 00:39:08.206511 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-interconnect-selfsigned-cert\" (UniqueName: \"kubernetes.io/secret/7298274a-5d57-47f7-b2a7-49869fc2f21c-default-interconnect-selfsigned-cert\") pod \"qdr-test\" (UID: \"7298274a-5d57-47f7-b2a7-49869fc2f21c\") " pod="service-telemetry/qdr-test" Dec 03 00:39:08 crc kubenswrapper[4953]: I1203 00:39:08.227574 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vqcb2\" (UniqueName: \"kubernetes.io/projected/7298274a-5d57-47f7-b2a7-49869fc2f21c-kube-api-access-vqcb2\") pod \"qdr-test\" (UID: \"7298274a-5d57-47f7-b2a7-49869fc2f21c\") " pod="service-telemetry/qdr-test" Dec 03 00:39:08 crc kubenswrapper[4953]: I1203 00:39:08.348408 4953 scope.go:117] "RemoveContainer" containerID="150ffc325f215c6da77fc72304c13cb10dafc800bbc932afd319cc34442a3f76" Dec 03 00:39:08 crc kubenswrapper[4953]: E1203 00:39:08.348633 4953 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6dfbn_openshift-machine-config-operator(53ffb1e9-9a2e-4869-80d7-776f4aa7ef8b)\"" pod="openshift-machine-config-operator/machine-config-daemon-6dfbn" podUID="53ffb1e9-9a2e-4869-80d7-776f4aa7ef8b" Dec 03 00:39:08 crc kubenswrapper[4953]: I1203 00:39:08.368917 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/qdr-test" Dec 03 00:39:08 crc kubenswrapper[4953]: I1203 00:39:08.863570 4953 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/qdr-test"] Dec 03 00:39:09 crc kubenswrapper[4953]: I1203 00:39:09.379535 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/qdr-test" event={"ID":"7298274a-5d57-47f7-b2a7-49869fc2f21c","Type":"ContainerStarted","Data":"5d3fff36e780440f0dfd81dee2942cff6f22f340c7642479f2285323f3474c30"} Dec 03 00:39:13 crc kubenswrapper[4953]: I1203 00:39:13.348602 4953 scope.go:117] "RemoveContainer" containerID="d7b8179b1510c185c7e7579faefd9679e1cc2a3085b4dcddaa45bb459a502caa" Dec 03 00:39:14 crc kubenswrapper[4953]: I1203 00:39:14.348105 4953 scope.go:117] "RemoveContainer" containerID="8459c45dddae86d9f97b4b43679ae5e69556a32c8d91a72b2d96967724932e45" Dec 03 00:39:15 crc kubenswrapper[4953]: I1203 00:39:15.354307 4953 scope.go:117] "RemoveContainer" containerID="64e62958118a8c50351495310c184e1c9a0003e66823530ee93a94a3a0bbf56f" Dec 03 00:39:16 crc kubenswrapper[4953]: I1203 00:39:16.348503 4953 scope.go:117] "RemoveContainer" containerID="22c8df2d89ef32be5e4546e57460efdeffc11511a7493e4df9aa89d10a537710" Dec 03 00:39:16 crc kubenswrapper[4953]: I1203 00:39:16.349163 4953 scope.go:117] "RemoveContainer" containerID="eeb6f10a572352c31fe4dc1894037f7459c6ff2ea7d5601fe9cecb9084034f7d" Dec 03 00:39:20 crc kubenswrapper[4953]: I1203 00:39:20.500176 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-57948895dc-blgcn" event={"ID":"43d17ba9-6fdf-4dca-991c-924f350daa0b","Type":"ContainerStarted","Data":"d0fd7b99e88016718182822624c6c3562e7dea9af980594f4d68962e4842e1f2"} Dec 03 00:39:20 crc kubenswrapper[4953]: I1203 00:39:20.503426 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/qdr-test" event={"ID":"7298274a-5d57-47f7-b2a7-49869fc2f21c","Type":"ContainerStarted","Data":"6effda0a2c05ec49d870c9a1d44d83940b0a3cf1007809935dad48738061805c"} Dec 03 00:39:20 crc kubenswrapper[4953]: I1203 00:39:20.507475 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-coll-meter-smartgateway-7cd87f9766-bgbxc" event={"ID":"f2ba70b7-7676-47af-aa77-4f90579e7ec1","Type":"ContainerStarted","Data":"bf0e75ed931ac5563f4ec2ce2bb0e1466d3197762f72c8c9fd4265c5ea9217fa"} Dec 03 00:39:20 crc kubenswrapper[4953]: I1203 00:39:20.511899 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-sens-meter-smartgateway-5759b4d97-nlgdm" event={"ID":"d660753e-48eb-4388-bf97-c7050b9cb1d2","Type":"ContainerStarted","Data":"4146eb489171092b7f1370eabd3d99dfc8d4b10e222a05d12bf6b76a88ca082b"} Dec 03 00:39:20 crc kubenswrapper[4953]: I1203 00:39:20.519505 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-ceil-event-smartgateway-5967d7cb84-dgpqt" event={"ID":"b9fb7d41-3caf-41bc-ac33-23d6757b4b11","Type":"ContainerStarted","Data":"8c7dcca443735494cd482748a756369f75ca746a535a77bf52fb6cdf08ed7aad"} Dec 03 00:39:20 crc kubenswrapper[4953]: I1203 00:39:20.527305 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-coll-event-smartgateway-5bf48db5df-92twf" event={"ID":"9683141e-460a-4680-81ca-58907eb72789","Type":"ContainerStarted","Data":"90776523a6d2358db58e64ec587a140f931a845b8a55e42206512ad5edd4389f"} Dec 03 00:39:20 crc kubenswrapper[4953]: I1203 00:39:20.627651 4953 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="service-telemetry/qdr-test" podStartSLOduration=1.8595722430000001 podStartE2EDuration="12.627620184s" podCreationTimestamp="2025-12-03 00:39:08 +0000 UTC" firstStartedPulling="2025-12-03 00:39:08.87973795 +0000 UTC m=+1937.864018733" lastFinishedPulling="2025-12-03 00:39:19.647785891 +0000 UTC m=+1948.632066674" observedRunningTime="2025-12-03 00:39:20.622759574 +0000 UTC m=+1949.607040367" watchObservedRunningTime="2025-12-03 00:39:20.627620184 +0000 UTC m=+1949.611900967" Dec 03 00:39:20 crc kubenswrapper[4953]: I1203 00:39:20.959121 4953 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["service-telemetry/stf-smoketest-smoke1-6jklg"] Dec 03 00:39:20 crc kubenswrapper[4953]: I1203 00:39:20.960276 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/stf-smoketest-smoke1-6jklg" Dec 03 00:39:20 crc kubenswrapper[4953]: I1203 00:39:20.962233 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"stf-smoketest-collectd-entrypoint-script" Dec 03 00:39:20 crc kubenswrapper[4953]: I1203 00:39:20.963017 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"stf-smoketest-sensubility-config" Dec 03 00:39:20 crc kubenswrapper[4953]: I1203 00:39:20.963280 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"stf-smoketest-healthcheck-log" Dec 03 00:39:20 crc kubenswrapper[4953]: I1203 00:39:20.963666 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"stf-smoketest-ceilometer-entrypoint-script" Dec 03 00:39:20 crc kubenswrapper[4953]: I1203 00:39:20.963863 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"stf-smoketest-ceilometer-publisher" Dec 03 00:39:20 crc kubenswrapper[4953]: I1203 00:39:20.978669 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"stf-smoketest-collectd-config" Dec 03 00:39:20 crc kubenswrapper[4953]: I1203 00:39:20.987408 4953 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/stf-smoketest-smoke1-6jklg"] Dec 03 00:39:21 crc kubenswrapper[4953]: I1203 00:39:21.137819 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"collectd-entrypoint-script\" (UniqueName: \"kubernetes.io/configmap/8cf971bf-230e-47dc-b5ca-73f8d36bbed0-collectd-entrypoint-script\") pod \"stf-smoketest-smoke1-6jklg\" (UID: \"8cf971bf-230e-47dc-b5ca-73f8d36bbed0\") " pod="service-telemetry/stf-smoketest-smoke1-6jklg" Dec 03 00:39:21 crc kubenswrapper[4953]: I1203 00:39:21.137897 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-entrypoint-script\" (UniqueName: \"kubernetes.io/configmap/8cf971bf-230e-47dc-b5ca-73f8d36bbed0-ceilometer-entrypoint-script\") pod \"stf-smoketest-smoke1-6jklg\" (UID: \"8cf971bf-230e-47dc-b5ca-73f8d36bbed0\") " pod="service-telemetry/stf-smoketest-smoke1-6jklg" Dec 03 00:39:21 crc kubenswrapper[4953]: I1203 00:39:21.137984 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m648c\" (UniqueName: \"kubernetes.io/projected/8cf971bf-230e-47dc-b5ca-73f8d36bbed0-kube-api-access-m648c\") pod \"stf-smoketest-smoke1-6jklg\" (UID: \"8cf971bf-230e-47dc-b5ca-73f8d36bbed0\") " pod="service-telemetry/stf-smoketest-smoke1-6jklg" Dec 03 00:39:21 crc kubenswrapper[4953]: I1203 00:39:21.138144 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"collectd-config\" (UniqueName: \"kubernetes.io/configmap/8cf971bf-230e-47dc-b5ca-73f8d36bbed0-collectd-config\") pod \"stf-smoketest-smoke1-6jklg\" (UID: \"8cf971bf-230e-47dc-b5ca-73f8d36bbed0\") " pod="service-telemetry/stf-smoketest-smoke1-6jklg" Dec 03 00:39:21 crc kubenswrapper[4953]: I1203 00:39:21.138252 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-publisher\" (UniqueName: \"kubernetes.io/configmap/8cf971bf-230e-47dc-b5ca-73f8d36bbed0-ceilometer-publisher\") pod \"stf-smoketest-smoke1-6jklg\" (UID: \"8cf971bf-230e-47dc-b5ca-73f8d36bbed0\") " pod="service-telemetry/stf-smoketest-smoke1-6jklg" Dec 03 00:39:21 crc kubenswrapper[4953]: I1203 00:39:21.138300 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sensubility-config\" (UniqueName: \"kubernetes.io/configmap/8cf971bf-230e-47dc-b5ca-73f8d36bbed0-sensubility-config\") pod \"stf-smoketest-smoke1-6jklg\" (UID: \"8cf971bf-230e-47dc-b5ca-73f8d36bbed0\") " pod="service-telemetry/stf-smoketest-smoke1-6jklg" Dec 03 00:39:21 crc kubenswrapper[4953]: I1203 00:39:21.138340 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"healthcheck-log\" (UniqueName: \"kubernetes.io/configmap/8cf971bf-230e-47dc-b5ca-73f8d36bbed0-healthcheck-log\") pod \"stf-smoketest-smoke1-6jklg\" (UID: \"8cf971bf-230e-47dc-b5ca-73f8d36bbed0\") " pod="service-telemetry/stf-smoketest-smoke1-6jklg" Dec 03 00:39:21 crc kubenswrapper[4953]: I1203 00:39:21.240477 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"collectd-entrypoint-script\" (UniqueName: \"kubernetes.io/configmap/8cf971bf-230e-47dc-b5ca-73f8d36bbed0-collectd-entrypoint-script\") pod \"stf-smoketest-smoke1-6jklg\" (UID: \"8cf971bf-230e-47dc-b5ca-73f8d36bbed0\") " pod="service-telemetry/stf-smoketest-smoke1-6jklg" Dec 03 00:39:21 crc kubenswrapper[4953]: I1203 00:39:21.240586 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-entrypoint-script\" (UniqueName: \"kubernetes.io/configmap/8cf971bf-230e-47dc-b5ca-73f8d36bbed0-ceilometer-entrypoint-script\") pod \"stf-smoketest-smoke1-6jklg\" (UID: \"8cf971bf-230e-47dc-b5ca-73f8d36bbed0\") " pod="service-telemetry/stf-smoketest-smoke1-6jklg" Dec 03 00:39:21 crc kubenswrapper[4953]: I1203 00:39:21.240656 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m648c\" (UniqueName: \"kubernetes.io/projected/8cf971bf-230e-47dc-b5ca-73f8d36bbed0-kube-api-access-m648c\") pod \"stf-smoketest-smoke1-6jklg\" (UID: \"8cf971bf-230e-47dc-b5ca-73f8d36bbed0\") " pod="service-telemetry/stf-smoketest-smoke1-6jklg" Dec 03 00:39:21 crc kubenswrapper[4953]: I1203 00:39:21.240701 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"collectd-config\" (UniqueName: \"kubernetes.io/configmap/8cf971bf-230e-47dc-b5ca-73f8d36bbed0-collectd-config\") pod \"stf-smoketest-smoke1-6jklg\" (UID: \"8cf971bf-230e-47dc-b5ca-73f8d36bbed0\") " pod="service-telemetry/stf-smoketest-smoke1-6jklg" Dec 03 00:39:21 crc kubenswrapper[4953]: I1203 00:39:21.240770 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-publisher\" (UniqueName: \"kubernetes.io/configmap/8cf971bf-230e-47dc-b5ca-73f8d36bbed0-ceilometer-publisher\") pod \"stf-smoketest-smoke1-6jklg\" (UID: \"8cf971bf-230e-47dc-b5ca-73f8d36bbed0\") " pod="service-telemetry/stf-smoketest-smoke1-6jklg" Dec 03 00:39:21 crc kubenswrapper[4953]: I1203 00:39:21.240815 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sensubility-config\" (UniqueName: \"kubernetes.io/configmap/8cf971bf-230e-47dc-b5ca-73f8d36bbed0-sensubility-config\") pod \"stf-smoketest-smoke1-6jklg\" (UID: \"8cf971bf-230e-47dc-b5ca-73f8d36bbed0\") " pod="service-telemetry/stf-smoketest-smoke1-6jklg" Dec 03 00:39:21 crc kubenswrapper[4953]: I1203 00:39:21.240857 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"healthcheck-log\" (UniqueName: \"kubernetes.io/configmap/8cf971bf-230e-47dc-b5ca-73f8d36bbed0-healthcheck-log\") pod \"stf-smoketest-smoke1-6jklg\" (UID: \"8cf971bf-230e-47dc-b5ca-73f8d36bbed0\") " pod="service-telemetry/stf-smoketest-smoke1-6jklg" Dec 03 00:39:21 crc kubenswrapper[4953]: I1203 00:39:21.241936 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"collectd-config\" (UniqueName: \"kubernetes.io/configmap/8cf971bf-230e-47dc-b5ca-73f8d36bbed0-collectd-config\") pod \"stf-smoketest-smoke1-6jklg\" (UID: \"8cf971bf-230e-47dc-b5ca-73f8d36bbed0\") " pod="service-telemetry/stf-smoketest-smoke1-6jklg" Dec 03 00:39:21 crc kubenswrapper[4953]: I1203 00:39:21.241996 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"healthcheck-log\" (UniqueName: \"kubernetes.io/configmap/8cf971bf-230e-47dc-b5ca-73f8d36bbed0-healthcheck-log\") pod \"stf-smoketest-smoke1-6jklg\" (UID: \"8cf971bf-230e-47dc-b5ca-73f8d36bbed0\") " pod="service-telemetry/stf-smoketest-smoke1-6jklg" Dec 03 00:39:21 crc kubenswrapper[4953]: I1203 00:39:21.243020 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-publisher\" (UniqueName: \"kubernetes.io/configmap/8cf971bf-230e-47dc-b5ca-73f8d36bbed0-ceilometer-publisher\") pod \"stf-smoketest-smoke1-6jklg\" (UID: \"8cf971bf-230e-47dc-b5ca-73f8d36bbed0\") " pod="service-telemetry/stf-smoketest-smoke1-6jklg" Dec 03 00:39:21 crc kubenswrapper[4953]: I1203 00:39:21.243317 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"collectd-entrypoint-script\" (UniqueName: \"kubernetes.io/configmap/8cf971bf-230e-47dc-b5ca-73f8d36bbed0-collectd-entrypoint-script\") pod \"stf-smoketest-smoke1-6jklg\" (UID: \"8cf971bf-230e-47dc-b5ca-73f8d36bbed0\") " pod="service-telemetry/stf-smoketest-smoke1-6jklg" Dec 03 00:39:21 crc kubenswrapper[4953]: I1203 00:39:21.244108 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sensubility-config\" (UniqueName: \"kubernetes.io/configmap/8cf971bf-230e-47dc-b5ca-73f8d36bbed0-sensubility-config\") pod \"stf-smoketest-smoke1-6jklg\" (UID: \"8cf971bf-230e-47dc-b5ca-73f8d36bbed0\") " pod="service-telemetry/stf-smoketest-smoke1-6jklg" Dec 03 00:39:21 crc kubenswrapper[4953]: I1203 00:39:21.244965 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-entrypoint-script\" (UniqueName: \"kubernetes.io/configmap/8cf971bf-230e-47dc-b5ca-73f8d36bbed0-ceilometer-entrypoint-script\") pod \"stf-smoketest-smoke1-6jklg\" (UID: \"8cf971bf-230e-47dc-b5ca-73f8d36bbed0\") " pod="service-telemetry/stf-smoketest-smoke1-6jklg" Dec 03 00:39:21 crc kubenswrapper[4953]: I1203 00:39:21.284808 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m648c\" (UniqueName: \"kubernetes.io/projected/8cf971bf-230e-47dc-b5ca-73f8d36bbed0-kube-api-access-m648c\") pod \"stf-smoketest-smoke1-6jklg\" (UID: \"8cf971bf-230e-47dc-b5ca-73f8d36bbed0\") " pod="service-telemetry/stf-smoketest-smoke1-6jklg" Dec 03 00:39:21 crc kubenswrapper[4953]: I1203 00:39:21.367951 4953 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["service-telemetry/curl"] Dec 03 00:39:21 crc kubenswrapper[4953]: I1203 00:39:21.369301 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/curl" Dec 03 00:39:21 crc kubenswrapper[4953]: I1203 00:39:21.380658 4953 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/curl"] Dec 03 00:39:21 crc kubenswrapper[4953]: I1203 00:39:21.444659 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hc84t\" (UniqueName: \"kubernetes.io/projected/0273a5a9-73b9-4064-8cf2-304a7776181f-kube-api-access-hc84t\") pod \"curl\" (UID: \"0273a5a9-73b9-4064-8cf2-304a7776181f\") " pod="service-telemetry/curl" Dec 03 00:39:21 crc kubenswrapper[4953]: I1203 00:39:21.546486 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hc84t\" (UniqueName: \"kubernetes.io/projected/0273a5a9-73b9-4064-8cf2-304a7776181f-kube-api-access-hc84t\") pod \"curl\" (UID: \"0273a5a9-73b9-4064-8cf2-304a7776181f\") " pod="service-telemetry/curl" Dec 03 00:39:21 crc kubenswrapper[4953]: I1203 00:39:21.570643 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hc84t\" (UniqueName: \"kubernetes.io/projected/0273a5a9-73b9-4064-8cf2-304a7776181f-kube-api-access-hc84t\") pod \"curl\" (UID: \"0273a5a9-73b9-4064-8cf2-304a7776181f\") " pod="service-telemetry/curl" Dec 03 00:39:21 crc kubenswrapper[4953]: I1203 00:39:21.579505 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/stf-smoketest-smoke1-6jklg" Dec 03 00:39:21 crc kubenswrapper[4953]: I1203 00:39:21.695659 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/curl" Dec 03 00:39:21 crc kubenswrapper[4953]: I1203 00:39:21.884117 4953 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/stf-smoketest-smoke1-6jklg"] Dec 03 00:39:22 crc kubenswrapper[4953]: I1203 00:39:22.193188 4953 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/curl"] Dec 03 00:39:22 crc kubenswrapper[4953]: I1203 00:39:22.347522 4953 scope.go:117] "RemoveContainer" containerID="150ffc325f215c6da77fc72304c13cb10dafc800bbc932afd319cc34442a3f76" Dec 03 00:39:22 crc kubenswrapper[4953]: I1203 00:39:22.548352 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/stf-smoketest-smoke1-6jklg" event={"ID":"8cf971bf-230e-47dc-b5ca-73f8d36bbed0","Type":"ContainerStarted","Data":"bb9ad370548525a8203d208936f625a5e619be8e6e5cb0aa2e0614bac07c5407"} Dec 03 00:39:22 crc kubenswrapper[4953]: I1203 00:39:22.549416 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/curl" event={"ID":"0273a5a9-73b9-4064-8cf2-304a7776181f","Type":"ContainerStarted","Data":"0e113e520d1a95ed9c3062e0edfa2cc4f31c596b3a566371e4ae31a0226670a4"} Dec 03 00:39:23 crc kubenswrapper[4953]: I1203 00:39:23.578627 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-6dfbn" event={"ID":"53ffb1e9-9a2e-4869-80d7-776f4aa7ef8b","Type":"ContainerStarted","Data":"e9bb3f6087811f1ea895def278b13c2a3d71dcdf55ad22dee5da1051917891f2"} Dec 03 00:39:25 crc kubenswrapper[4953]: I1203 00:39:25.600470 4953 generic.go:334] "Generic (PLEG): container finished" podID="0273a5a9-73b9-4064-8cf2-304a7776181f" containerID="c397fb6a418639c2acd9d6275ec9c59bbb953700faba62e480f0efae908a42f4" exitCode=0 Dec 03 00:39:25 crc kubenswrapper[4953]: I1203 00:39:25.600591 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/curl" event={"ID":"0273a5a9-73b9-4064-8cf2-304a7776181f","Type":"ContainerDied","Data":"c397fb6a418639c2acd9d6275ec9c59bbb953700faba62e480f0efae908a42f4"} Dec 03 00:39:34 crc kubenswrapper[4953]: I1203 00:39:34.658532 4953 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/curl" Dec 03 00:39:34 crc kubenswrapper[4953]: I1203 00:39:34.712847 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/curl" event={"ID":"0273a5a9-73b9-4064-8cf2-304a7776181f","Type":"ContainerDied","Data":"0e113e520d1a95ed9c3062e0edfa2cc4f31c596b3a566371e4ae31a0226670a4"} Dec 03 00:39:34 crc kubenswrapper[4953]: I1203 00:39:34.712906 4953 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0e113e520d1a95ed9c3062e0edfa2cc4f31c596b3a566371e4ae31a0226670a4" Dec 03 00:39:34 crc kubenswrapper[4953]: I1203 00:39:34.713042 4953 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/curl" Dec 03 00:39:34 crc kubenswrapper[4953]: I1203 00:39:34.826996 4953 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_curl_0273a5a9-73b9-4064-8cf2-304a7776181f/curl/0.log" Dec 03 00:39:34 crc kubenswrapper[4953]: I1203 00:39:34.843701 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hc84t\" (UniqueName: \"kubernetes.io/projected/0273a5a9-73b9-4064-8cf2-304a7776181f-kube-api-access-hc84t\") pod \"0273a5a9-73b9-4064-8cf2-304a7776181f\" (UID: \"0273a5a9-73b9-4064-8cf2-304a7776181f\") " Dec 03 00:39:34 crc kubenswrapper[4953]: I1203 00:39:34.849738 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0273a5a9-73b9-4064-8cf2-304a7776181f-kube-api-access-hc84t" (OuterVolumeSpecName: "kube-api-access-hc84t") pod "0273a5a9-73b9-4064-8cf2-304a7776181f" (UID: "0273a5a9-73b9-4064-8cf2-304a7776181f"). InnerVolumeSpecName "kube-api-access-hc84t". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:39:34 crc kubenswrapper[4953]: I1203 00:39:34.946403 4953 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hc84t\" (UniqueName: \"kubernetes.io/projected/0273a5a9-73b9-4064-8cf2-304a7776181f-kube-api-access-hc84t\") on node \"crc\" DevicePath \"\"" Dec 03 00:39:35 crc kubenswrapper[4953]: I1203 00:39:35.142213 4953 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_default-snmp-webhook-6856cfb745-jf6rj_7bf37579-93eb-42d1-91ed-2ea71e4683ae/prometheus-webhook-snmp/0.log" Dec 03 00:39:35 crc kubenswrapper[4953]: I1203 00:39:35.728947 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/stf-smoketest-smoke1-6jklg" event={"ID":"8cf971bf-230e-47dc-b5ca-73f8d36bbed0","Type":"ContainerStarted","Data":"da1b7b065cd8778fdd240cf6c9d57115d3711aa2e24b0e7bcda69ee4cb2093f6"} Dec 03 00:39:42 crc kubenswrapper[4953]: I1203 00:39:42.808935 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/stf-smoketest-smoke1-6jklg" event={"ID":"8cf971bf-230e-47dc-b5ca-73f8d36bbed0","Type":"ContainerStarted","Data":"9d5d403a9c00d06ee9aff7390aa8d5a7cdd9c43a01ceeb4fae434bfeee8fc8e0"} Dec 03 00:39:42 crc kubenswrapper[4953]: I1203 00:39:42.842090 4953 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="service-telemetry/stf-smoketest-smoke1-6jklg" podStartSLOduration=2.624809279 podStartE2EDuration="22.842066978s" podCreationTimestamp="2025-12-03 00:39:20 +0000 UTC" firstStartedPulling="2025-12-03 00:39:21.905753931 +0000 UTC m=+1950.890034714" lastFinishedPulling="2025-12-03 00:39:42.12301163 +0000 UTC m=+1971.107292413" observedRunningTime="2025-12-03 00:39:42.840441398 +0000 UTC m=+1971.824722221" watchObservedRunningTime="2025-12-03 00:39:42.842066978 +0000 UTC m=+1971.826347761" Dec 03 00:40:05 crc kubenswrapper[4953]: I1203 00:40:05.303163 4953 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_default-snmp-webhook-6856cfb745-jf6rj_7bf37579-93eb-42d1-91ed-2ea71e4683ae/prometheus-webhook-snmp/0.log" Dec 03 00:40:09 crc kubenswrapper[4953]: I1203 00:40:09.157307 4953 generic.go:334] "Generic (PLEG): container finished" podID="8cf971bf-230e-47dc-b5ca-73f8d36bbed0" containerID="da1b7b065cd8778fdd240cf6c9d57115d3711aa2e24b0e7bcda69ee4cb2093f6" exitCode=0 Dec 03 00:40:09 crc kubenswrapper[4953]: I1203 00:40:09.157391 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/stf-smoketest-smoke1-6jklg" event={"ID":"8cf971bf-230e-47dc-b5ca-73f8d36bbed0","Type":"ContainerDied","Data":"da1b7b065cd8778fdd240cf6c9d57115d3711aa2e24b0e7bcda69ee4cb2093f6"} Dec 03 00:40:09 crc kubenswrapper[4953]: I1203 00:40:09.159277 4953 scope.go:117] "RemoveContainer" containerID="da1b7b065cd8778fdd240cf6c9d57115d3711aa2e24b0e7bcda69ee4cb2093f6" Dec 03 00:40:15 crc kubenswrapper[4953]: I1203 00:40:15.221654 4953 generic.go:334] "Generic (PLEG): container finished" podID="8cf971bf-230e-47dc-b5ca-73f8d36bbed0" containerID="9d5d403a9c00d06ee9aff7390aa8d5a7cdd9c43a01ceeb4fae434bfeee8fc8e0" exitCode=0 Dec 03 00:40:15 crc kubenswrapper[4953]: I1203 00:40:15.221880 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/stf-smoketest-smoke1-6jklg" event={"ID":"8cf971bf-230e-47dc-b5ca-73f8d36bbed0","Type":"ContainerDied","Data":"9d5d403a9c00d06ee9aff7390aa8d5a7cdd9c43a01ceeb4fae434bfeee8fc8e0"} Dec 03 00:40:16 crc kubenswrapper[4953]: I1203 00:40:16.536631 4953 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/stf-smoketest-smoke1-6jklg" Dec 03 00:40:16 crc kubenswrapper[4953]: I1203 00:40:16.621719 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"healthcheck-log\" (UniqueName: \"kubernetes.io/configmap/8cf971bf-230e-47dc-b5ca-73f8d36bbed0-healthcheck-log\") pod \"8cf971bf-230e-47dc-b5ca-73f8d36bbed0\" (UID: \"8cf971bf-230e-47dc-b5ca-73f8d36bbed0\") " Dec 03 00:40:16 crc kubenswrapper[4953]: I1203 00:40:16.621830 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"collectd-config\" (UniqueName: \"kubernetes.io/configmap/8cf971bf-230e-47dc-b5ca-73f8d36bbed0-collectd-config\") pod \"8cf971bf-230e-47dc-b5ca-73f8d36bbed0\" (UID: \"8cf971bf-230e-47dc-b5ca-73f8d36bbed0\") " Dec 03 00:40:16 crc kubenswrapper[4953]: I1203 00:40:16.642715 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8cf971bf-230e-47dc-b5ca-73f8d36bbed0-collectd-config" (OuterVolumeSpecName: "collectd-config") pod "8cf971bf-230e-47dc-b5ca-73f8d36bbed0" (UID: "8cf971bf-230e-47dc-b5ca-73f8d36bbed0"). InnerVolumeSpecName "collectd-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:40:16 crc kubenswrapper[4953]: I1203 00:40:16.647931 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8cf971bf-230e-47dc-b5ca-73f8d36bbed0-healthcheck-log" (OuterVolumeSpecName: "healthcheck-log") pod "8cf971bf-230e-47dc-b5ca-73f8d36bbed0" (UID: "8cf971bf-230e-47dc-b5ca-73f8d36bbed0"). InnerVolumeSpecName "healthcheck-log". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:40:16 crc kubenswrapper[4953]: I1203 00:40:16.723263 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"collectd-entrypoint-script\" (UniqueName: \"kubernetes.io/configmap/8cf971bf-230e-47dc-b5ca-73f8d36bbed0-collectd-entrypoint-script\") pod \"8cf971bf-230e-47dc-b5ca-73f8d36bbed0\" (UID: \"8cf971bf-230e-47dc-b5ca-73f8d36bbed0\") " Dec 03 00:40:16 crc kubenswrapper[4953]: I1203 00:40:16.723355 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m648c\" (UniqueName: \"kubernetes.io/projected/8cf971bf-230e-47dc-b5ca-73f8d36bbed0-kube-api-access-m648c\") pod \"8cf971bf-230e-47dc-b5ca-73f8d36bbed0\" (UID: \"8cf971bf-230e-47dc-b5ca-73f8d36bbed0\") " Dec 03 00:40:16 crc kubenswrapper[4953]: I1203 00:40:16.723399 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sensubility-config\" (UniqueName: \"kubernetes.io/configmap/8cf971bf-230e-47dc-b5ca-73f8d36bbed0-sensubility-config\") pod \"8cf971bf-230e-47dc-b5ca-73f8d36bbed0\" (UID: \"8cf971bf-230e-47dc-b5ca-73f8d36bbed0\") " Dec 03 00:40:16 crc kubenswrapper[4953]: I1203 00:40:16.723449 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-entrypoint-script\" (UniqueName: \"kubernetes.io/configmap/8cf971bf-230e-47dc-b5ca-73f8d36bbed0-ceilometer-entrypoint-script\") pod \"8cf971bf-230e-47dc-b5ca-73f8d36bbed0\" (UID: \"8cf971bf-230e-47dc-b5ca-73f8d36bbed0\") " Dec 03 00:40:16 crc kubenswrapper[4953]: I1203 00:40:16.723482 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-publisher\" (UniqueName: \"kubernetes.io/configmap/8cf971bf-230e-47dc-b5ca-73f8d36bbed0-ceilometer-publisher\") pod \"8cf971bf-230e-47dc-b5ca-73f8d36bbed0\" (UID: \"8cf971bf-230e-47dc-b5ca-73f8d36bbed0\") " Dec 03 00:40:16 crc kubenswrapper[4953]: I1203 00:40:16.724490 4953 reconciler_common.go:293] "Volume detached for volume \"healthcheck-log\" (UniqueName: \"kubernetes.io/configmap/8cf971bf-230e-47dc-b5ca-73f8d36bbed0-healthcheck-log\") on node \"crc\" DevicePath \"\"" Dec 03 00:40:16 crc kubenswrapper[4953]: I1203 00:40:16.724528 4953 reconciler_common.go:293] "Volume detached for volume \"collectd-config\" (UniqueName: \"kubernetes.io/configmap/8cf971bf-230e-47dc-b5ca-73f8d36bbed0-collectd-config\") on node \"crc\" DevicePath \"\"" Dec 03 00:40:16 crc kubenswrapper[4953]: I1203 00:40:16.728305 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8cf971bf-230e-47dc-b5ca-73f8d36bbed0-kube-api-access-m648c" (OuterVolumeSpecName: "kube-api-access-m648c") pod "8cf971bf-230e-47dc-b5ca-73f8d36bbed0" (UID: "8cf971bf-230e-47dc-b5ca-73f8d36bbed0"). InnerVolumeSpecName "kube-api-access-m648c". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:40:16 crc kubenswrapper[4953]: I1203 00:40:16.741698 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8cf971bf-230e-47dc-b5ca-73f8d36bbed0-sensubility-config" (OuterVolumeSpecName: "sensubility-config") pod "8cf971bf-230e-47dc-b5ca-73f8d36bbed0" (UID: "8cf971bf-230e-47dc-b5ca-73f8d36bbed0"). InnerVolumeSpecName "sensubility-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:40:16 crc kubenswrapper[4953]: I1203 00:40:16.742734 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8cf971bf-230e-47dc-b5ca-73f8d36bbed0-collectd-entrypoint-script" (OuterVolumeSpecName: "collectd-entrypoint-script") pod "8cf971bf-230e-47dc-b5ca-73f8d36bbed0" (UID: "8cf971bf-230e-47dc-b5ca-73f8d36bbed0"). InnerVolumeSpecName "collectd-entrypoint-script". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:40:16 crc kubenswrapper[4953]: E1203 00:40:16.744295 4953 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/8cf971bf-230e-47dc-b5ca-73f8d36bbed0-ceilometer-publisher podName:8cf971bf-230e-47dc-b5ca-73f8d36bbed0 nodeName:}" failed. No retries permitted until 2025-12-03 00:40:17.244261336 +0000 UTC m=+2006.228542119 (durationBeforeRetry 500ms). Error: error cleaning subPath mounts for volume "ceilometer-publisher" (UniqueName: "kubernetes.io/configmap/8cf971bf-230e-47dc-b5ca-73f8d36bbed0-ceilometer-publisher") pod "8cf971bf-230e-47dc-b5ca-73f8d36bbed0" (UID: "8cf971bf-230e-47dc-b5ca-73f8d36bbed0") : error deleting /var/lib/kubelet/pods/8cf971bf-230e-47dc-b5ca-73f8d36bbed0/volume-subpaths: remove /var/lib/kubelet/pods/8cf971bf-230e-47dc-b5ca-73f8d36bbed0/volume-subpaths: no such file or directory Dec 03 00:40:16 crc kubenswrapper[4953]: I1203 00:40:16.744704 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8cf971bf-230e-47dc-b5ca-73f8d36bbed0-ceilometer-entrypoint-script" (OuterVolumeSpecName: "ceilometer-entrypoint-script") pod "8cf971bf-230e-47dc-b5ca-73f8d36bbed0" (UID: "8cf971bf-230e-47dc-b5ca-73f8d36bbed0"). InnerVolumeSpecName "ceilometer-entrypoint-script". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:40:16 crc kubenswrapper[4953]: I1203 00:40:16.825992 4953 reconciler_common.go:293] "Volume detached for volume \"collectd-entrypoint-script\" (UniqueName: \"kubernetes.io/configmap/8cf971bf-230e-47dc-b5ca-73f8d36bbed0-collectd-entrypoint-script\") on node \"crc\" DevicePath \"\"" Dec 03 00:40:16 crc kubenswrapper[4953]: I1203 00:40:16.826044 4953 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m648c\" (UniqueName: \"kubernetes.io/projected/8cf971bf-230e-47dc-b5ca-73f8d36bbed0-kube-api-access-m648c\") on node \"crc\" DevicePath \"\"" Dec 03 00:40:16 crc kubenswrapper[4953]: I1203 00:40:16.826058 4953 reconciler_common.go:293] "Volume detached for volume \"sensubility-config\" (UniqueName: \"kubernetes.io/configmap/8cf971bf-230e-47dc-b5ca-73f8d36bbed0-sensubility-config\") on node \"crc\" DevicePath \"\"" Dec 03 00:40:16 crc kubenswrapper[4953]: I1203 00:40:16.826069 4953 reconciler_common.go:293] "Volume detached for volume \"ceilometer-entrypoint-script\" (UniqueName: \"kubernetes.io/configmap/8cf971bf-230e-47dc-b5ca-73f8d36bbed0-ceilometer-entrypoint-script\") on node \"crc\" DevicePath \"\"" Dec 03 00:40:17 crc kubenswrapper[4953]: I1203 00:40:17.244727 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/stf-smoketest-smoke1-6jklg" event={"ID":"8cf971bf-230e-47dc-b5ca-73f8d36bbed0","Type":"ContainerDied","Data":"bb9ad370548525a8203d208936f625a5e619be8e6e5cb0aa2e0614bac07c5407"} Dec 03 00:40:17 crc kubenswrapper[4953]: I1203 00:40:17.244790 4953 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bb9ad370548525a8203d208936f625a5e619be8e6e5cb0aa2e0614bac07c5407" Dec 03 00:40:17 crc kubenswrapper[4953]: I1203 00:40:17.244919 4953 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/stf-smoketest-smoke1-6jklg" Dec 03 00:40:17 crc kubenswrapper[4953]: I1203 00:40:17.336020 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-publisher\" (UniqueName: \"kubernetes.io/configmap/8cf971bf-230e-47dc-b5ca-73f8d36bbed0-ceilometer-publisher\") pod \"8cf971bf-230e-47dc-b5ca-73f8d36bbed0\" (UID: \"8cf971bf-230e-47dc-b5ca-73f8d36bbed0\") " Dec 03 00:40:17 crc kubenswrapper[4953]: I1203 00:40:17.336799 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8cf971bf-230e-47dc-b5ca-73f8d36bbed0-ceilometer-publisher" (OuterVolumeSpecName: "ceilometer-publisher") pod "8cf971bf-230e-47dc-b5ca-73f8d36bbed0" (UID: "8cf971bf-230e-47dc-b5ca-73f8d36bbed0"). InnerVolumeSpecName "ceilometer-publisher". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:40:17 crc kubenswrapper[4953]: I1203 00:40:17.438369 4953 reconciler_common.go:293] "Volume detached for volume \"ceilometer-publisher\" (UniqueName: \"kubernetes.io/configmap/8cf971bf-230e-47dc-b5ca-73f8d36bbed0-ceilometer-publisher\") on node \"crc\" DevicePath \"\"" Dec 03 00:40:18 crc kubenswrapper[4953]: I1203 00:40:18.589068 4953 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_stf-smoketest-smoke1-6jklg_8cf971bf-230e-47dc-b5ca-73f8d36bbed0/smoketest-collectd/0.log" Dec 03 00:40:18 crc kubenswrapper[4953]: I1203 00:40:18.829726 4953 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_stf-smoketest-smoke1-6jklg_8cf971bf-230e-47dc-b5ca-73f8d36bbed0/smoketest-ceilometer/0.log" Dec 03 00:40:19 crc kubenswrapper[4953]: I1203 00:40:19.191500 4953 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_default-interconnect-68864d46cb-4644b_122f9ca6-e8e4-45ac-bfc4-0eec579dee72/default-interconnect/0.log" Dec 03 00:40:19 crc kubenswrapper[4953]: I1203 00:40:19.552935 4953 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_default-cloud1-coll-meter-smartgateway-7cd87f9766-bgbxc_f2ba70b7-7676-47af-aa77-4f90579e7ec1/bridge/2.log" Dec 03 00:40:19 crc kubenswrapper[4953]: I1203 00:40:19.905082 4953 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_default-cloud1-coll-meter-smartgateway-7cd87f9766-bgbxc_f2ba70b7-7676-47af-aa77-4f90579e7ec1/sg-core/0.log" Dec 03 00:40:20 crc kubenswrapper[4953]: I1203 00:40:20.191784 4953 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_default-cloud1-coll-event-smartgateway-5bf48db5df-92twf_9683141e-460a-4680-81ca-58907eb72789/bridge/2.log" Dec 03 00:40:20 crc kubenswrapper[4953]: I1203 00:40:20.587882 4953 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_default-cloud1-coll-event-smartgateway-5bf48db5df-92twf_9683141e-460a-4680-81ca-58907eb72789/sg-core/0.log" Dec 03 00:40:20 crc kubenswrapper[4953]: I1203 00:40:20.850296 4953 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_default-cloud1-ceil-meter-smartgateway-57948895dc-blgcn_43d17ba9-6fdf-4dca-991c-924f350daa0b/bridge/2.log" Dec 03 00:40:21 crc kubenswrapper[4953]: I1203 00:40:21.146152 4953 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_default-cloud1-ceil-meter-smartgateway-57948895dc-blgcn_43d17ba9-6fdf-4dca-991c-924f350daa0b/sg-core/0.log" Dec 03 00:40:21 crc kubenswrapper[4953]: I1203 00:40:21.412624 4953 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_default-cloud1-ceil-event-smartgateway-5967d7cb84-dgpqt_b9fb7d41-3caf-41bc-ac33-23d6757b4b11/bridge/2.log" Dec 03 00:40:21 crc kubenswrapper[4953]: I1203 00:40:21.648831 4953 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_default-cloud1-ceil-event-smartgateway-5967d7cb84-dgpqt_b9fb7d41-3caf-41bc-ac33-23d6757b4b11/sg-core/0.log" Dec 03 00:40:21 crc kubenswrapper[4953]: I1203 00:40:21.942927 4953 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_default-cloud1-sens-meter-smartgateway-5759b4d97-nlgdm_d660753e-48eb-4388-bf97-c7050b9cb1d2/bridge/2.log" Dec 03 00:40:22 crc kubenswrapper[4953]: I1203 00:40:22.214966 4953 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_default-cloud1-sens-meter-smartgateway-5759b4d97-nlgdm_d660753e-48eb-4388-bf97-c7050b9cb1d2/sg-core/0.log" Dec 03 00:40:25 crc kubenswrapper[4953]: I1203 00:40:25.625921 4953 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_smart-gateway-operator-7cddd7d4d-vnx98_34c1166e-eb61-4a66-82bd-6498b4d908a8/operator/0.log" Dec 03 00:40:25 crc kubenswrapper[4953]: I1203 00:40:25.893311 4953 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_prometheus-default-0_603c47f8-e2e2-4dbd-aa98-a0eca1acf697/prometheus/0.log" Dec 03 00:40:26 crc kubenswrapper[4953]: I1203 00:40:26.267523 4953 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_elasticsearch-es-default-0_3313dc1f-c9a1-4617-a9d6-cdf983a9d618/elasticsearch/0.log" Dec 03 00:40:26 crc kubenswrapper[4953]: I1203 00:40:26.630051 4953 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_default-snmp-webhook-6856cfb745-jf6rj_7bf37579-93eb-42d1-91ed-2ea71e4683ae/prometheus-webhook-snmp/0.log" Dec 03 00:40:26 crc kubenswrapper[4953]: I1203 00:40:26.968649 4953 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_alertmanager-default-0_1df63057-30c5-4625-8365-9fe9907156c9/alertmanager/0.log" Dec 03 00:40:42 crc kubenswrapper[4953]: I1203 00:40:42.490968 4953 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_service-telemetry-operator-75f4bf7844-lfwk2_8353057a-2770-4006-a75c-37c59f034192/operator/0.log" Dec 03 00:40:45 crc kubenswrapper[4953]: I1203 00:40:45.939006 4953 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_smart-gateway-operator-7cddd7d4d-vnx98_34c1166e-eb61-4a66-82bd-6498b4d908a8/operator/0.log" Dec 03 00:40:46 crc kubenswrapper[4953]: I1203 00:40:46.196314 4953 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_qdr-test_7298274a-5d57-47f7-b2a7-49869fc2f21c/qdr/0.log" Dec 03 00:40:50 crc kubenswrapper[4953]: I1203 00:40:50.040119 4953 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["service-telemetry/infrawatch-operators-mj65f"] Dec 03 00:40:50 crc kubenswrapper[4953]: E1203 00:40:50.041159 4953 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8cf971bf-230e-47dc-b5ca-73f8d36bbed0" containerName="smoketest-ceilometer" Dec 03 00:40:50 crc kubenswrapper[4953]: I1203 00:40:50.041175 4953 state_mem.go:107] "Deleted CPUSet assignment" podUID="8cf971bf-230e-47dc-b5ca-73f8d36bbed0" containerName="smoketest-ceilometer" Dec 03 00:40:50 crc kubenswrapper[4953]: E1203 00:40:50.041187 4953 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0273a5a9-73b9-4064-8cf2-304a7776181f" containerName="curl" Dec 03 00:40:50 crc kubenswrapper[4953]: I1203 00:40:50.041194 4953 state_mem.go:107] "Deleted CPUSet assignment" podUID="0273a5a9-73b9-4064-8cf2-304a7776181f" containerName="curl" Dec 03 00:40:50 crc kubenswrapper[4953]: E1203 00:40:50.041221 4953 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8cf971bf-230e-47dc-b5ca-73f8d36bbed0" containerName="smoketest-collectd" Dec 03 00:40:50 crc kubenswrapper[4953]: I1203 00:40:50.041229 4953 state_mem.go:107] "Deleted CPUSet assignment" podUID="8cf971bf-230e-47dc-b5ca-73f8d36bbed0" containerName="smoketest-collectd" Dec 03 00:40:50 crc kubenswrapper[4953]: I1203 00:40:50.041383 4953 memory_manager.go:354] "RemoveStaleState removing state" podUID="8cf971bf-230e-47dc-b5ca-73f8d36bbed0" containerName="smoketest-collectd" Dec 03 00:40:50 crc kubenswrapper[4953]: I1203 00:40:50.041394 4953 memory_manager.go:354] "RemoveStaleState removing state" podUID="0273a5a9-73b9-4064-8cf2-304a7776181f" containerName="curl" Dec 03 00:40:50 crc kubenswrapper[4953]: I1203 00:40:50.041412 4953 memory_manager.go:354] "RemoveStaleState removing state" podUID="8cf971bf-230e-47dc-b5ca-73f8d36bbed0" containerName="smoketest-ceilometer" Dec 03 00:40:50 crc kubenswrapper[4953]: I1203 00:40:50.041985 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/infrawatch-operators-mj65f" Dec 03 00:40:50 crc kubenswrapper[4953]: I1203 00:40:50.043960 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wzhks\" (UniqueName: \"kubernetes.io/projected/cc85cd70-142f-4b16-9cd0-c583292d8cc7-kube-api-access-wzhks\") pod \"infrawatch-operators-mj65f\" (UID: \"cc85cd70-142f-4b16-9cd0-c583292d8cc7\") " pod="service-telemetry/infrawatch-operators-mj65f" Dec 03 00:40:50 crc kubenswrapper[4953]: I1203 00:40:50.054957 4953 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/infrawatch-operators-mj65f"] Dec 03 00:40:50 crc kubenswrapper[4953]: I1203 00:40:50.145293 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wzhks\" (UniqueName: \"kubernetes.io/projected/cc85cd70-142f-4b16-9cd0-c583292d8cc7-kube-api-access-wzhks\") pod \"infrawatch-operators-mj65f\" (UID: \"cc85cd70-142f-4b16-9cd0-c583292d8cc7\") " pod="service-telemetry/infrawatch-operators-mj65f" Dec 03 00:40:50 crc kubenswrapper[4953]: I1203 00:40:50.166784 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wzhks\" (UniqueName: \"kubernetes.io/projected/cc85cd70-142f-4b16-9cd0-c583292d8cc7-kube-api-access-wzhks\") pod \"infrawatch-operators-mj65f\" (UID: \"cc85cd70-142f-4b16-9cd0-c583292d8cc7\") " pod="service-telemetry/infrawatch-operators-mj65f" Dec 03 00:40:50 crc kubenswrapper[4953]: I1203 00:40:50.371483 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/infrawatch-operators-mj65f" Dec 03 00:40:51 crc kubenswrapper[4953]: I1203 00:40:51.250591 4953 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/infrawatch-operators-mj65f"] Dec 03 00:40:51 crc kubenswrapper[4953]: I1203 00:40:51.532778 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/infrawatch-operators-mj65f" event={"ID":"cc85cd70-142f-4b16-9cd0-c583292d8cc7","Type":"ContainerStarted","Data":"721e948ea7357b5f1ccacbdbb55a9d020fa8cf14da388f376389825e6f2cd959"} Dec 03 00:40:51 crc kubenswrapper[4953]: I1203 00:40:51.533185 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/infrawatch-operators-mj65f" event={"ID":"cc85cd70-142f-4b16-9cd0-c583292d8cc7","Type":"ContainerStarted","Data":"5723ece85cf78f568d5039076de96910a9eda2a45760831b1b2ea76e6d455d37"} Dec 03 00:40:51 crc kubenswrapper[4953]: I1203 00:40:51.554007 4953 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="service-telemetry/infrawatch-operators-mj65f" podStartSLOduration=1.430903702 podStartE2EDuration="1.553986909s" podCreationTimestamp="2025-12-03 00:40:50 +0000 UTC" firstStartedPulling="2025-12-03 00:40:51.260073594 +0000 UTC m=+2040.244354387" lastFinishedPulling="2025-12-03 00:40:51.383156811 +0000 UTC m=+2040.367437594" observedRunningTime="2025-12-03 00:40:51.546374651 +0000 UTC m=+2040.530655434" watchObservedRunningTime="2025-12-03 00:40:51.553986909 +0000 UTC m=+2040.538267692" Dec 03 00:40:52 crc kubenswrapper[4953]: I1203 00:40:52.626873 4953 scope.go:117] "RemoveContainer" containerID="2dc6fee63c189a01beee6babd1ef39aac49ee77e0084c0f2e4a36d2bc9f18d9c" Dec 03 00:40:52 crc kubenswrapper[4953]: I1203 00:40:52.655789 4953 scope.go:117] "RemoveContainer" containerID="c3f17a1d41fe0ce559e682dccbc4fcd8027a2761d9c8f4187ff4eb27a32e90b7" Dec 03 00:40:52 crc kubenswrapper[4953]: I1203 00:40:52.688232 4953 scope.go:117] "RemoveContainer" containerID="c6561219aec20a1d2fe783d1c29fe9039b37748e18a43f060ad617826ad78c43" Dec 03 00:40:52 crc kubenswrapper[4953]: I1203 00:40:52.725738 4953 scope.go:117] "RemoveContainer" containerID="dd54eea3f4a21825be55d64ec038a1e721a99ec10166dd9fd1ef4a28c3b9c664" Dec 03 00:41:00 crc kubenswrapper[4953]: I1203 00:41:00.372680 4953 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="service-telemetry/infrawatch-operators-mj65f" Dec 03 00:41:00 crc kubenswrapper[4953]: I1203 00:41:00.373204 4953 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="service-telemetry/infrawatch-operators-mj65f" Dec 03 00:41:00 crc kubenswrapper[4953]: I1203 00:41:00.420233 4953 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="service-telemetry/infrawatch-operators-mj65f" Dec 03 00:41:00 crc kubenswrapper[4953]: I1203 00:41:00.626761 4953 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="service-telemetry/infrawatch-operators-mj65f" Dec 03 00:41:00 crc kubenswrapper[4953]: I1203 00:41:00.666304 4953 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["service-telemetry/infrawatch-operators-mj65f"] Dec 03 00:41:02 crc kubenswrapper[4953]: I1203 00:41:02.619677 4953 kuberuntime_container.go:808] "Killing container with a grace period" pod="service-telemetry/infrawatch-operators-mj65f" podUID="cc85cd70-142f-4b16-9cd0-c583292d8cc7" containerName="registry-server" containerID="cri-o://721e948ea7357b5f1ccacbdbb55a9d020fa8cf14da388f376389825e6f2cd959" gracePeriod=2 Dec 03 00:41:03 crc kubenswrapper[4953]: I1203 00:41:03.103092 4953 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/infrawatch-operators-mj65f" Dec 03 00:41:03 crc kubenswrapper[4953]: I1203 00:41:03.245752 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wzhks\" (UniqueName: \"kubernetes.io/projected/cc85cd70-142f-4b16-9cd0-c583292d8cc7-kube-api-access-wzhks\") pod \"cc85cd70-142f-4b16-9cd0-c583292d8cc7\" (UID: \"cc85cd70-142f-4b16-9cd0-c583292d8cc7\") " Dec 03 00:41:03 crc kubenswrapper[4953]: I1203 00:41:03.251469 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cc85cd70-142f-4b16-9cd0-c583292d8cc7-kube-api-access-wzhks" (OuterVolumeSpecName: "kube-api-access-wzhks") pod "cc85cd70-142f-4b16-9cd0-c583292d8cc7" (UID: "cc85cd70-142f-4b16-9cd0-c583292d8cc7"). InnerVolumeSpecName "kube-api-access-wzhks". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:41:03 crc kubenswrapper[4953]: I1203 00:41:03.348078 4953 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wzhks\" (UniqueName: \"kubernetes.io/projected/cc85cd70-142f-4b16-9cd0-c583292d8cc7-kube-api-access-wzhks\") on node \"crc\" DevicePath \"\"" Dec 03 00:41:03 crc kubenswrapper[4953]: I1203 00:41:03.632532 4953 generic.go:334] "Generic (PLEG): container finished" podID="cc85cd70-142f-4b16-9cd0-c583292d8cc7" containerID="721e948ea7357b5f1ccacbdbb55a9d020fa8cf14da388f376389825e6f2cd959" exitCode=0 Dec 03 00:41:03 crc kubenswrapper[4953]: I1203 00:41:03.632706 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/infrawatch-operators-mj65f" event={"ID":"cc85cd70-142f-4b16-9cd0-c583292d8cc7","Type":"ContainerDied","Data":"721e948ea7357b5f1ccacbdbb55a9d020fa8cf14da388f376389825e6f2cd959"} Dec 03 00:41:03 crc kubenswrapper[4953]: I1203 00:41:03.632764 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/infrawatch-operators-mj65f" event={"ID":"cc85cd70-142f-4b16-9cd0-c583292d8cc7","Type":"ContainerDied","Data":"5723ece85cf78f568d5039076de96910a9eda2a45760831b1b2ea76e6d455d37"} Dec 03 00:41:03 crc kubenswrapper[4953]: I1203 00:41:03.632800 4953 scope.go:117] "RemoveContainer" containerID="721e948ea7357b5f1ccacbdbb55a9d020fa8cf14da388f376389825e6f2cd959" Dec 03 00:41:03 crc kubenswrapper[4953]: I1203 00:41:03.633036 4953 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/infrawatch-operators-mj65f" Dec 03 00:41:03 crc kubenswrapper[4953]: I1203 00:41:03.668192 4953 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["service-telemetry/infrawatch-operators-mj65f"] Dec 03 00:41:03 crc kubenswrapper[4953]: I1203 00:41:03.671327 4953 scope.go:117] "RemoveContainer" containerID="721e948ea7357b5f1ccacbdbb55a9d020fa8cf14da388f376389825e6f2cd959" Dec 03 00:41:03 crc kubenswrapper[4953]: E1203 00:41:03.671849 4953 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"721e948ea7357b5f1ccacbdbb55a9d020fa8cf14da388f376389825e6f2cd959\": container with ID starting with 721e948ea7357b5f1ccacbdbb55a9d020fa8cf14da388f376389825e6f2cd959 not found: ID does not exist" containerID="721e948ea7357b5f1ccacbdbb55a9d020fa8cf14da388f376389825e6f2cd959" Dec 03 00:41:03 crc kubenswrapper[4953]: I1203 00:41:03.671881 4953 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"721e948ea7357b5f1ccacbdbb55a9d020fa8cf14da388f376389825e6f2cd959"} err="failed to get container status \"721e948ea7357b5f1ccacbdbb55a9d020fa8cf14da388f376389825e6f2cd959\": rpc error: code = NotFound desc = could not find container \"721e948ea7357b5f1ccacbdbb55a9d020fa8cf14da388f376389825e6f2cd959\": container with ID starting with 721e948ea7357b5f1ccacbdbb55a9d020fa8cf14da388f376389825e6f2cd959 not found: ID does not exist" Dec 03 00:41:03 crc kubenswrapper[4953]: I1203 00:41:03.677771 4953 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["service-telemetry/infrawatch-operators-mj65f"] Dec 03 00:41:05 crc kubenswrapper[4953]: I1203 00:41:05.358028 4953 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cc85cd70-142f-4b16-9cd0-c583292d8cc7" path="/var/lib/kubelet/pods/cc85cd70-142f-4b16-9cd0-c583292d8cc7/volumes" Dec 03 00:41:10 crc kubenswrapper[4953]: I1203 00:41:10.772146 4953 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-mnrv5/must-gather-s69cf"] Dec 03 00:41:10 crc kubenswrapper[4953]: E1203 00:41:10.773048 4953 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cc85cd70-142f-4b16-9cd0-c583292d8cc7" containerName="registry-server" Dec 03 00:41:10 crc kubenswrapper[4953]: I1203 00:41:10.773072 4953 state_mem.go:107] "Deleted CPUSet assignment" podUID="cc85cd70-142f-4b16-9cd0-c583292d8cc7" containerName="registry-server" Dec 03 00:41:10 crc kubenswrapper[4953]: I1203 00:41:10.773269 4953 memory_manager.go:354] "RemoveStaleState removing state" podUID="cc85cd70-142f-4b16-9cd0-c583292d8cc7" containerName="registry-server" Dec 03 00:41:10 crc kubenswrapper[4953]: I1203 00:41:10.774440 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-mnrv5/must-gather-s69cf" Dec 03 00:41:10 crc kubenswrapper[4953]: I1203 00:41:10.778824 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-mnrv5"/"openshift-service-ca.crt" Dec 03 00:41:10 crc kubenswrapper[4953]: I1203 00:41:10.778911 4953 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-mnrv5"/"default-dockercfg-xmv9h" Dec 03 00:41:10 crc kubenswrapper[4953]: I1203 00:41:10.779256 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-mnrv5"/"kube-root-ca.crt" Dec 03 00:41:10 crc kubenswrapper[4953]: I1203 00:41:10.796825 4953 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-mnrv5/must-gather-s69cf"] Dec 03 00:41:10 crc kubenswrapper[4953]: I1203 00:41:10.879798 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/7ee12f05-7ff1-4efa-b45e-45376c77b906-must-gather-output\") pod \"must-gather-s69cf\" (UID: \"7ee12f05-7ff1-4efa-b45e-45376c77b906\") " pod="openshift-must-gather-mnrv5/must-gather-s69cf" Dec 03 00:41:10 crc kubenswrapper[4953]: I1203 00:41:10.880259 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jtc9v\" (UniqueName: \"kubernetes.io/projected/7ee12f05-7ff1-4efa-b45e-45376c77b906-kube-api-access-jtc9v\") pod \"must-gather-s69cf\" (UID: \"7ee12f05-7ff1-4efa-b45e-45376c77b906\") " pod="openshift-must-gather-mnrv5/must-gather-s69cf" Dec 03 00:41:10 crc kubenswrapper[4953]: I1203 00:41:10.982354 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jtc9v\" (UniqueName: \"kubernetes.io/projected/7ee12f05-7ff1-4efa-b45e-45376c77b906-kube-api-access-jtc9v\") pod \"must-gather-s69cf\" (UID: \"7ee12f05-7ff1-4efa-b45e-45376c77b906\") " pod="openshift-must-gather-mnrv5/must-gather-s69cf" Dec 03 00:41:10 crc kubenswrapper[4953]: I1203 00:41:10.983354 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/7ee12f05-7ff1-4efa-b45e-45376c77b906-must-gather-output\") pod \"must-gather-s69cf\" (UID: \"7ee12f05-7ff1-4efa-b45e-45376c77b906\") " pod="openshift-must-gather-mnrv5/must-gather-s69cf" Dec 03 00:41:10 crc kubenswrapper[4953]: I1203 00:41:10.983753 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/7ee12f05-7ff1-4efa-b45e-45376c77b906-must-gather-output\") pod \"must-gather-s69cf\" (UID: \"7ee12f05-7ff1-4efa-b45e-45376c77b906\") " pod="openshift-must-gather-mnrv5/must-gather-s69cf" Dec 03 00:41:11 crc kubenswrapper[4953]: I1203 00:41:11.009153 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jtc9v\" (UniqueName: \"kubernetes.io/projected/7ee12f05-7ff1-4efa-b45e-45376c77b906-kube-api-access-jtc9v\") pod \"must-gather-s69cf\" (UID: \"7ee12f05-7ff1-4efa-b45e-45376c77b906\") " pod="openshift-must-gather-mnrv5/must-gather-s69cf" Dec 03 00:41:11 crc kubenswrapper[4953]: I1203 00:41:11.094886 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-mnrv5/must-gather-s69cf" Dec 03 00:41:11 crc kubenswrapper[4953]: I1203 00:41:11.550656 4953 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-mnrv5/must-gather-s69cf"] Dec 03 00:41:11 crc kubenswrapper[4953]: I1203 00:41:11.721929 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-mnrv5/must-gather-s69cf" event={"ID":"7ee12f05-7ff1-4efa-b45e-45376c77b906","Type":"ContainerStarted","Data":"6c8d10d81bf159675837503c6ef8e2f11732580c7f7875d5f09a85ff0765fff9"} Dec 03 00:41:21 crc kubenswrapper[4953]: I1203 00:41:21.817497 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-mnrv5/must-gather-s69cf" event={"ID":"7ee12f05-7ff1-4efa-b45e-45376c77b906","Type":"ContainerStarted","Data":"a778f07eebaf40f3618ceb12ade247a3b45a9bfabd835e0b96f52a8325be6ba2"} Dec 03 00:41:22 crc kubenswrapper[4953]: I1203 00:41:22.827843 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-mnrv5/must-gather-s69cf" event={"ID":"7ee12f05-7ff1-4efa-b45e-45376c77b906","Type":"ContainerStarted","Data":"02447eb8bbe00c7b77ea6b8275874ea50738776b27034961751231aebfe8efc1"} Dec 03 00:41:22 crc kubenswrapper[4953]: I1203 00:41:22.851150 4953 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-mnrv5/must-gather-s69cf" podStartSLOduration=3.038534975 podStartE2EDuration="12.851122656s" podCreationTimestamp="2025-12-03 00:41:10 +0000 UTC" firstStartedPulling="2025-12-03 00:41:11.566852648 +0000 UTC m=+2060.551133441" lastFinishedPulling="2025-12-03 00:41:21.379440339 +0000 UTC m=+2070.363721122" observedRunningTime="2025-12-03 00:41:22.846666236 +0000 UTC m=+2071.830947019" watchObservedRunningTime="2025-12-03 00:41:22.851122656 +0000 UTC m=+2071.835403429" Dec 03 00:41:48 crc kubenswrapper[4953]: I1203 00:41:48.945206 4953 patch_prober.go:28] interesting pod/machine-config-daemon-6dfbn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 00:41:48 crc kubenswrapper[4953]: I1203 00:41:48.945926 4953 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-6dfbn" podUID="53ffb1e9-9a2e-4869-80d7-776f4aa7ef8b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 00:42:09 crc kubenswrapper[4953]: I1203 00:42:09.150423 4953 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-xhc8f_4da5668c-a644-4014-8358-790da0f313e2/control-plane-machine-set-operator/0.log" Dec 03 00:42:09 crc kubenswrapper[4953]: I1203 00:42:09.232816 4953 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-xbwlm_496793fd-c6ab-48e4-a2c5-c259593c5d1c/kube-rbac-proxy/0.log" Dec 03 00:42:09 crc kubenswrapper[4953]: I1203 00:42:09.381859 4953 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-xbwlm_496793fd-c6ab-48e4-a2c5-c259593c5d1c/machine-api-operator/0.log" Dec 03 00:42:18 crc kubenswrapper[4953]: I1203 00:42:18.945060 4953 patch_prober.go:28] interesting pod/machine-config-daemon-6dfbn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 00:42:18 crc kubenswrapper[4953]: I1203 00:42:18.945887 4953 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-6dfbn" podUID="53ffb1e9-9a2e-4869-80d7-776f4aa7ef8b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 00:42:23 crc kubenswrapper[4953]: I1203 00:42:23.505860 4953 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-86cb77c54b-hq8kc_d6305126-b142-45e0-a058-1356b122cecd/cert-manager-controller/0.log" Dec 03 00:42:23 crc kubenswrapper[4953]: I1203 00:42:23.761781 4953 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-855d9ccff4-gqhrh_e2c5f086-eaf3-49ae-b67c-e3eb601c2c58/cert-manager-cainjector/0.log" Dec 03 00:42:23 crc kubenswrapper[4953]: I1203 00:42:23.835384 4953 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-f4fb5df64-cg7mx_37a3ced5-065d-4be8-a3e9-54deee67aefa/cert-manager-webhook/0.log" Dec 03 00:42:41 crc kubenswrapper[4953]: I1203 00:42:41.215275 4953 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931ahqk5h_8bc0877b-282d-457a-a46a-c3202a31e2fb/util/0.log" Dec 03 00:42:41 crc kubenswrapper[4953]: I1203 00:42:41.437084 4953 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931ahqk5h_8bc0877b-282d-457a-a46a-c3202a31e2fb/util/0.log" Dec 03 00:42:41 crc kubenswrapper[4953]: I1203 00:42:41.468877 4953 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931ahqk5h_8bc0877b-282d-457a-a46a-c3202a31e2fb/pull/0.log" Dec 03 00:42:41 crc kubenswrapper[4953]: I1203 00:42:41.538145 4953 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931ahqk5h_8bc0877b-282d-457a-a46a-c3202a31e2fb/pull/0.log" Dec 03 00:42:41 crc kubenswrapper[4953]: I1203 00:42:41.611573 4953 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931ahqk5h_8bc0877b-282d-457a-a46a-c3202a31e2fb/util/0.log" Dec 03 00:42:41 crc kubenswrapper[4953]: I1203 00:42:41.697073 4953 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931ahqk5h_8bc0877b-282d-457a-a46a-c3202a31e2fb/pull/0.log" Dec 03 00:42:41 crc kubenswrapper[4953]: I1203 00:42:41.709573 4953 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931ahqk5h_8bc0877b-282d-457a-a46a-c3202a31e2fb/extract/0.log" Dec 03 00:42:41 crc kubenswrapper[4953]: I1203 00:42:41.850368 4953 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210kb6ph_474c987a-6ccd-4d96-bfb1-ae4cb96751c6/util/0.log" Dec 03 00:42:42 crc kubenswrapper[4953]: I1203 00:42:42.019697 4953 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210kb6ph_474c987a-6ccd-4d96-bfb1-ae4cb96751c6/pull/0.log" Dec 03 00:42:42 crc kubenswrapper[4953]: I1203 00:42:42.019856 4953 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210kb6ph_474c987a-6ccd-4d96-bfb1-ae4cb96751c6/pull/0.log" Dec 03 00:42:42 crc kubenswrapper[4953]: I1203 00:42:42.055675 4953 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210kb6ph_474c987a-6ccd-4d96-bfb1-ae4cb96751c6/util/0.log" Dec 03 00:42:42 crc kubenswrapper[4953]: I1203 00:42:42.239538 4953 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210kb6ph_474c987a-6ccd-4d96-bfb1-ae4cb96751c6/util/0.log" Dec 03 00:42:42 crc kubenswrapper[4953]: I1203 00:42:42.240144 4953 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210kb6ph_474c987a-6ccd-4d96-bfb1-ae4cb96751c6/pull/0.log" Dec 03 00:42:42 crc kubenswrapper[4953]: I1203 00:42:42.240275 4953 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210kb6ph_474c987a-6ccd-4d96-bfb1-ae4cb96751c6/extract/0.log" Dec 03 00:42:42 crc kubenswrapper[4953]: I1203 00:42:42.452656 4953 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fj4nbn_76a42fdc-9500-45c7-b826-5697261a08b0/util/0.log" Dec 03 00:42:42 crc kubenswrapper[4953]: I1203 00:42:42.660344 4953 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fj4nbn_76a42fdc-9500-45c7-b826-5697261a08b0/pull/0.log" Dec 03 00:42:42 crc kubenswrapper[4953]: I1203 00:42:42.675716 4953 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fj4nbn_76a42fdc-9500-45c7-b826-5697261a08b0/util/0.log" Dec 03 00:42:42 crc kubenswrapper[4953]: I1203 00:42:42.700355 4953 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fj4nbn_76a42fdc-9500-45c7-b826-5697261a08b0/pull/0.log" Dec 03 00:42:42 crc kubenswrapper[4953]: I1203 00:42:42.858881 4953 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fj4nbn_76a42fdc-9500-45c7-b826-5697261a08b0/util/0.log" Dec 03 00:42:42 crc kubenswrapper[4953]: I1203 00:42:42.859057 4953 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fj4nbn_76a42fdc-9500-45c7-b826-5697261a08b0/extract/0.log" Dec 03 00:42:42 crc kubenswrapper[4953]: I1203 00:42:42.874831 4953 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fj4nbn_76a42fdc-9500-45c7-b826-5697261a08b0/pull/0.log" Dec 03 00:42:43 crc kubenswrapper[4953]: I1203 00:42:43.054171 4953 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8ed862a309935d5a1c8012df79b93f7fb46e029d4689f7f6ddcb9e7f5ejn8gr_315d7133-df43-414b-a2af-93c2534cf3ba/util/0.log" Dec 03 00:42:43 crc kubenswrapper[4953]: I1203 00:42:43.220374 4953 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8ed862a309935d5a1c8012df79b93f7fb46e029d4689f7f6ddcb9e7f5ejn8gr_315d7133-df43-414b-a2af-93c2534cf3ba/util/0.log" Dec 03 00:42:43 crc kubenswrapper[4953]: I1203 00:42:43.235994 4953 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8ed862a309935d5a1c8012df79b93f7fb46e029d4689f7f6ddcb9e7f5ejn8gr_315d7133-df43-414b-a2af-93c2534cf3ba/pull/0.log" Dec 03 00:42:43 crc kubenswrapper[4953]: I1203 00:42:43.236029 4953 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8ed862a309935d5a1c8012df79b93f7fb46e029d4689f7f6ddcb9e7f5ejn8gr_315d7133-df43-414b-a2af-93c2534cf3ba/pull/0.log" Dec 03 00:42:43 crc kubenswrapper[4953]: I1203 00:42:43.449124 4953 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8ed862a309935d5a1c8012df79b93f7fb46e029d4689f7f6ddcb9e7f5ejn8gr_315d7133-df43-414b-a2af-93c2534cf3ba/pull/0.log" Dec 03 00:42:43 crc kubenswrapper[4953]: I1203 00:42:43.474808 4953 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8ed862a309935d5a1c8012df79b93f7fb46e029d4689f7f6ddcb9e7f5ejn8gr_315d7133-df43-414b-a2af-93c2534cf3ba/util/0.log" Dec 03 00:42:43 crc kubenswrapper[4953]: I1203 00:42:43.523491 4953 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8ed862a309935d5a1c8012df79b93f7fb46e029d4689f7f6ddcb9e7f5ejn8gr_315d7133-df43-414b-a2af-93c2534cf3ba/extract/0.log" Dec 03 00:42:43 crc kubenswrapper[4953]: I1203 00:42:43.677088 4953 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-lcrlw_475ce63b-d94a-4b5f-b213-bbc2dd733e63/extract-utilities/0.log" Dec 03 00:42:43 crc kubenswrapper[4953]: I1203 00:42:43.882061 4953 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-lcrlw_475ce63b-d94a-4b5f-b213-bbc2dd733e63/extract-utilities/0.log" Dec 03 00:42:43 crc kubenswrapper[4953]: I1203 00:42:43.884038 4953 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-lcrlw_475ce63b-d94a-4b5f-b213-bbc2dd733e63/extract-content/0.log" Dec 03 00:42:43 crc kubenswrapper[4953]: I1203 00:42:43.895427 4953 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-lcrlw_475ce63b-d94a-4b5f-b213-bbc2dd733e63/extract-content/0.log" Dec 03 00:42:44 crc kubenswrapper[4953]: I1203 00:42:44.111233 4953 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-lcrlw_475ce63b-d94a-4b5f-b213-bbc2dd733e63/extract-content/0.log" Dec 03 00:42:44 crc kubenswrapper[4953]: I1203 00:42:44.147475 4953 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-lcrlw_475ce63b-d94a-4b5f-b213-bbc2dd733e63/extract-utilities/0.log" Dec 03 00:42:44 crc kubenswrapper[4953]: I1203 00:42:44.350493 4953 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-lxqn5_60145acc-39c4-483a-b635-f12937264ce0/extract-utilities/0.log" Dec 03 00:42:44 crc kubenswrapper[4953]: I1203 00:42:44.499471 4953 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-lcrlw_475ce63b-d94a-4b5f-b213-bbc2dd733e63/registry-server/0.log" Dec 03 00:42:44 crc kubenswrapper[4953]: I1203 00:42:44.573202 4953 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-lxqn5_60145acc-39c4-483a-b635-f12937264ce0/extract-utilities/0.log" Dec 03 00:42:44 crc kubenswrapper[4953]: I1203 00:42:44.610807 4953 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-lxqn5_60145acc-39c4-483a-b635-f12937264ce0/extract-content/0.log" Dec 03 00:42:44 crc kubenswrapper[4953]: I1203 00:42:44.662843 4953 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-lxqn5_60145acc-39c4-483a-b635-f12937264ce0/extract-content/0.log" Dec 03 00:42:44 crc kubenswrapper[4953]: I1203 00:42:44.837925 4953 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-lxqn5_60145acc-39c4-483a-b635-f12937264ce0/extract-utilities/0.log" Dec 03 00:42:44 crc kubenswrapper[4953]: I1203 00:42:44.874882 4953 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-lxqn5_60145acc-39c4-483a-b635-f12937264ce0/extract-content/0.log" Dec 03 00:42:44 crc kubenswrapper[4953]: I1203 00:42:44.947281 4953 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-d9khm_4ac535b1-a7ec-4999-aae4-d1c60c284aff/marketplace-operator/0.log" Dec 03 00:42:45 crc kubenswrapper[4953]: I1203 00:42:45.144351 4953 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-qrq2c_9b3f5bab-705f-4fa1-8f74-6ae6a6aac120/extract-utilities/0.log" Dec 03 00:42:45 crc kubenswrapper[4953]: I1203 00:42:45.226851 4953 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-lxqn5_60145acc-39c4-483a-b635-f12937264ce0/registry-server/0.log" Dec 03 00:42:45 crc kubenswrapper[4953]: I1203 00:42:45.344513 4953 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-qrq2c_9b3f5bab-705f-4fa1-8f74-6ae6a6aac120/extract-utilities/0.log" Dec 03 00:42:45 crc kubenswrapper[4953]: I1203 00:42:45.378474 4953 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-qrq2c_9b3f5bab-705f-4fa1-8f74-6ae6a6aac120/extract-content/0.log" Dec 03 00:42:45 crc kubenswrapper[4953]: I1203 00:42:45.410921 4953 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-qrq2c_9b3f5bab-705f-4fa1-8f74-6ae6a6aac120/extract-content/0.log" Dec 03 00:42:45 crc kubenswrapper[4953]: I1203 00:42:45.583006 4953 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-qrq2c_9b3f5bab-705f-4fa1-8f74-6ae6a6aac120/extract-utilities/0.log" Dec 03 00:42:45 crc kubenswrapper[4953]: I1203 00:42:45.589900 4953 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-qrq2c_9b3f5bab-705f-4fa1-8f74-6ae6a6aac120/extract-content/0.log" Dec 03 00:42:45 crc kubenswrapper[4953]: I1203 00:42:45.936070 4953 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-qrq2c_9b3f5bab-705f-4fa1-8f74-6ae6a6aac120/registry-server/0.log" Dec 03 00:42:48 crc kubenswrapper[4953]: I1203 00:42:48.945216 4953 patch_prober.go:28] interesting pod/machine-config-daemon-6dfbn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 00:42:48 crc kubenswrapper[4953]: I1203 00:42:48.945775 4953 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-6dfbn" podUID="53ffb1e9-9a2e-4869-80d7-776f4aa7ef8b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 00:42:48 crc kubenswrapper[4953]: I1203 00:42:48.945872 4953 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-6dfbn" Dec 03 00:42:48 crc kubenswrapper[4953]: I1203 00:42:48.947045 4953 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"e9bb3f6087811f1ea895def278b13c2a3d71dcdf55ad22dee5da1051917891f2"} pod="openshift-machine-config-operator/machine-config-daemon-6dfbn" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 03 00:42:48 crc kubenswrapper[4953]: I1203 00:42:48.947126 4953 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-6dfbn" podUID="53ffb1e9-9a2e-4869-80d7-776f4aa7ef8b" containerName="machine-config-daemon" containerID="cri-o://e9bb3f6087811f1ea895def278b13c2a3d71dcdf55ad22dee5da1051917891f2" gracePeriod=600 Dec 03 00:42:49 crc kubenswrapper[4953]: I1203 00:42:49.555702 4953 generic.go:334] "Generic (PLEG): container finished" podID="53ffb1e9-9a2e-4869-80d7-776f4aa7ef8b" containerID="e9bb3f6087811f1ea895def278b13c2a3d71dcdf55ad22dee5da1051917891f2" exitCode=0 Dec 03 00:42:49 crc kubenswrapper[4953]: I1203 00:42:49.555801 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-6dfbn" event={"ID":"53ffb1e9-9a2e-4869-80d7-776f4aa7ef8b","Type":"ContainerDied","Data":"e9bb3f6087811f1ea895def278b13c2a3d71dcdf55ad22dee5da1051917891f2"} Dec 03 00:42:49 crc kubenswrapper[4953]: I1203 00:42:49.556173 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-6dfbn" event={"ID":"53ffb1e9-9a2e-4869-80d7-776f4aa7ef8b","Type":"ContainerStarted","Data":"3aa2983089e4009bcb4ec8782554c6ba514ea720700589b3adbf3c56a194c241"} Dec 03 00:42:49 crc kubenswrapper[4953]: I1203 00:42:49.556210 4953 scope.go:117] "RemoveContainer" containerID="150ffc325f215c6da77fc72304c13cb10dafc800bbc932afd319cc34442a3f76" Dec 03 00:42:58 crc kubenswrapper[4953]: I1203 00:42:58.021778 4953 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-668cf9dfbb-kk8d4_f5d63f17-0a24-453c-a528-e61266f2b091/prometheus-operator/0.log" Dec 03 00:42:58 crc kubenswrapper[4953]: I1203 00:42:58.189662 4953 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-admission-webhook-68bbd5c68d-2lg5v_0ce8b8d2-06e2-4df3-9a3f-d2022263fd97/prometheus-operator-admission-webhook/0.log" Dec 03 00:42:58 crc kubenswrapper[4953]: I1203 00:42:58.269646 4953 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-admission-webhook-68bbd5c68d-pmttp_fc13989b-f854-49fd-8901-3b206167aafc/prometheus-operator-admission-webhook/0.log" Dec 03 00:42:58 crc kubenswrapper[4953]: I1203 00:42:58.427314 4953 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_observability-operator-d8bb48f5d-64xzt_1df642b8-586a-4bc5-8d9f-9f2cff7c6b80/operator/0.log" Dec 03 00:42:58 crc kubenswrapper[4953]: I1203 00:42:58.529898 4953 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_perses-operator-5446b9c989-hqmj8_61476fab-b1b9-424c-ab5d-853f83d6fdfd/perses-operator/0.log" Dec 03 00:43:38 crc kubenswrapper[4953]: I1203 00:43:38.775390 4953 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-r4xg5"] Dec 03 00:43:38 crc kubenswrapper[4953]: I1203 00:43:38.778638 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-r4xg5" Dec 03 00:43:38 crc kubenswrapper[4953]: I1203 00:43:38.801408 4953 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-r4xg5"] Dec 03 00:43:38 crc kubenswrapper[4953]: I1203 00:43:38.892676 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3bb49b8a-e497-48bf-bc83-5effc81b78e4-catalog-content\") pod \"redhat-operators-r4xg5\" (UID: \"3bb49b8a-e497-48bf-bc83-5effc81b78e4\") " pod="openshift-marketplace/redhat-operators-r4xg5" Dec 03 00:43:38 crc kubenswrapper[4953]: I1203 00:43:38.893155 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fmxmz\" (UniqueName: \"kubernetes.io/projected/3bb49b8a-e497-48bf-bc83-5effc81b78e4-kube-api-access-fmxmz\") pod \"redhat-operators-r4xg5\" (UID: \"3bb49b8a-e497-48bf-bc83-5effc81b78e4\") " pod="openshift-marketplace/redhat-operators-r4xg5" Dec 03 00:43:38 crc kubenswrapper[4953]: I1203 00:43:38.893320 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3bb49b8a-e497-48bf-bc83-5effc81b78e4-utilities\") pod \"redhat-operators-r4xg5\" (UID: \"3bb49b8a-e497-48bf-bc83-5effc81b78e4\") " pod="openshift-marketplace/redhat-operators-r4xg5" Dec 03 00:43:38 crc kubenswrapper[4953]: I1203 00:43:38.994857 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3bb49b8a-e497-48bf-bc83-5effc81b78e4-utilities\") pod \"redhat-operators-r4xg5\" (UID: \"3bb49b8a-e497-48bf-bc83-5effc81b78e4\") " pod="openshift-marketplace/redhat-operators-r4xg5" Dec 03 00:43:38 crc kubenswrapper[4953]: I1203 00:43:38.995003 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3bb49b8a-e497-48bf-bc83-5effc81b78e4-catalog-content\") pod \"redhat-operators-r4xg5\" (UID: \"3bb49b8a-e497-48bf-bc83-5effc81b78e4\") " pod="openshift-marketplace/redhat-operators-r4xg5" Dec 03 00:43:38 crc kubenswrapper[4953]: I1203 00:43:38.995060 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fmxmz\" (UniqueName: \"kubernetes.io/projected/3bb49b8a-e497-48bf-bc83-5effc81b78e4-kube-api-access-fmxmz\") pod \"redhat-operators-r4xg5\" (UID: \"3bb49b8a-e497-48bf-bc83-5effc81b78e4\") " pod="openshift-marketplace/redhat-operators-r4xg5" Dec 03 00:43:38 crc kubenswrapper[4953]: I1203 00:43:38.995700 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3bb49b8a-e497-48bf-bc83-5effc81b78e4-utilities\") pod \"redhat-operators-r4xg5\" (UID: \"3bb49b8a-e497-48bf-bc83-5effc81b78e4\") " pod="openshift-marketplace/redhat-operators-r4xg5" Dec 03 00:43:38 crc kubenswrapper[4953]: I1203 00:43:38.995741 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3bb49b8a-e497-48bf-bc83-5effc81b78e4-catalog-content\") pod \"redhat-operators-r4xg5\" (UID: \"3bb49b8a-e497-48bf-bc83-5effc81b78e4\") " pod="openshift-marketplace/redhat-operators-r4xg5" Dec 03 00:43:39 crc kubenswrapper[4953]: I1203 00:43:39.018194 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fmxmz\" (UniqueName: \"kubernetes.io/projected/3bb49b8a-e497-48bf-bc83-5effc81b78e4-kube-api-access-fmxmz\") pod \"redhat-operators-r4xg5\" (UID: \"3bb49b8a-e497-48bf-bc83-5effc81b78e4\") " pod="openshift-marketplace/redhat-operators-r4xg5" Dec 03 00:43:39 crc kubenswrapper[4953]: I1203 00:43:39.105376 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-r4xg5" Dec 03 00:43:39 crc kubenswrapper[4953]: I1203 00:43:39.616714 4953 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-r4xg5"] Dec 03 00:43:40 crc kubenswrapper[4953]: I1203 00:43:40.026883 4953 generic.go:334] "Generic (PLEG): container finished" podID="3bb49b8a-e497-48bf-bc83-5effc81b78e4" containerID="485392e1c2b391f689c44dc861da19005b55bcb9d5c594ba3c31a2eed7897ad1" exitCode=0 Dec 03 00:43:40 crc kubenswrapper[4953]: I1203 00:43:40.027056 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-r4xg5" event={"ID":"3bb49b8a-e497-48bf-bc83-5effc81b78e4","Type":"ContainerDied","Data":"485392e1c2b391f689c44dc861da19005b55bcb9d5c594ba3c31a2eed7897ad1"} Dec 03 00:43:40 crc kubenswrapper[4953]: I1203 00:43:40.027371 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-r4xg5" event={"ID":"3bb49b8a-e497-48bf-bc83-5effc81b78e4","Type":"ContainerStarted","Data":"1840c550eaa1c549084678e0b215d09bd1589b9282eb2e544c0224cee3f46434"} Dec 03 00:43:40 crc kubenswrapper[4953]: I1203 00:43:40.029869 4953 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 03 00:43:40 crc kubenswrapper[4953]: I1203 00:43:40.552213 4953 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-jtckq"] Dec 03 00:43:40 crc kubenswrapper[4953]: I1203 00:43:40.553826 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-jtckq" Dec 03 00:43:40 crc kubenswrapper[4953]: I1203 00:43:40.571499 4953 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-jtckq"] Dec 03 00:43:40 crc kubenswrapper[4953]: I1203 00:43:40.652516 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8f4dbfe6-01d5-43dc-9f8b-a1f78aa2397b-utilities\") pod \"certified-operators-jtckq\" (UID: \"8f4dbfe6-01d5-43dc-9f8b-a1f78aa2397b\") " pod="openshift-marketplace/certified-operators-jtckq" Dec 03 00:43:40 crc kubenswrapper[4953]: I1203 00:43:40.652687 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8f4dbfe6-01d5-43dc-9f8b-a1f78aa2397b-catalog-content\") pod \"certified-operators-jtckq\" (UID: \"8f4dbfe6-01d5-43dc-9f8b-a1f78aa2397b\") " pod="openshift-marketplace/certified-operators-jtckq" Dec 03 00:43:40 crc kubenswrapper[4953]: I1203 00:43:40.652738 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wzgdj\" (UniqueName: \"kubernetes.io/projected/8f4dbfe6-01d5-43dc-9f8b-a1f78aa2397b-kube-api-access-wzgdj\") pod \"certified-operators-jtckq\" (UID: \"8f4dbfe6-01d5-43dc-9f8b-a1f78aa2397b\") " pod="openshift-marketplace/certified-operators-jtckq" Dec 03 00:43:40 crc kubenswrapper[4953]: I1203 00:43:40.753936 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8f4dbfe6-01d5-43dc-9f8b-a1f78aa2397b-catalog-content\") pod \"certified-operators-jtckq\" (UID: \"8f4dbfe6-01d5-43dc-9f8b-a1f78aa2397b\") " pod="openshift-marketplace/certified-operators-jtckq" Dec 03 00:43:40 crc kubenswrapper[4953]: I1203 00:43:40.754022 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wzgdj\" (UniqueName: \"kubernetes.io/projected/8f4dbfe6-01d5-43dc-9f8b-a1f78aa2397b-kube-api-access-wzgdj\") pod \"certified-operators-jtckq\" (UID: \"8f4dbfe6-01d5-43dc-9f8b-a1f78aa2397b\") " pod="openshift-marketplace/certified-operators-jtckq" Dec 03 00:43:40 crc kubenswrapper[4953]: I1203 00:43:40.754124 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8f4dbfe6-01d5-43dc-9f8b-a1f78aa2397b-utilities\") pod \"certified-operators-jtckq\" (UID: \"8f4dbfe6-01d5-43dc-9f8b-a1f78aa2397b\") " pod="openshift-marketplace/certified-operators-jtckq" Dec 03 00:43:40 crc kubenswrapper[4953]: I1203 00:43:40.754716 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8f4dbfe6-01d5-43dc-9f8b-a1f78aa2397b-utilities\") pod \"certified-operators-jtckq\" (UID: \"8f4dbfe6-01d5-43dc-9f8b-a1f78aa2397b\") " pod="openshift-marketplace/certified-operators-jtckq" Dec 03 00:43:40 crc kubenswrapper[4953]: I1203 00:43:40.754716 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8f4dbfe6-01d5-43dc-9f8b-a1f78aa2397b-catalog-content\") pod \"certified-operators-jtckq\" (UID: \"8f4dbfe6-01d5-43dc-9f8b-a1f78aa2397b\") " pod="openshift-marketplace/certified-operators-jtckq" Dec 03 00:43:40 crc kubenswrapper[4953]: I1203 00:43:40.791047 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wzgdj\" (UniqueName: \"kubernetes.io/projected/8f4dbfe6-01d5-43dc-9f8b-a1f78aa2397b-kube-api-access-wzgdj\") pod \"certified-operators-jtckq\" (UID: \"8f4dbfe6-01d5-43dc-9f8b-a1f78aa2397b\") " pod="openshift-marketplace/certified-operators-jtckq" Dec 03 00:43:40 crc kubenswrapper[4953]: I1203 00:43:40.872533 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-jtckq" Dec 03 00:43:41 crc kubenswrapper[4953]: I1203 00:43:41.041714 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-r4xg5" event={"ID":"3bb49b8a-e497-48bf-bc83-5effc81b78e4","Type":"ContainerStarted","Data":"50f5cfa6b05f2d2ca54ad7d5d26191e72f9584044d74a4af470fd241779d348b"} Dec 03 00:43:41 crc kubenswrapper[4953]: I1203 00:43:41.413836 4953 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-jtckq"] Dec 03 00:43:41 crc kubenswrapper[4953]: W1203 00:43:41.419025 4953 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8f4dbfe6_01d5_43dc_9f8b_a1f78aa2397b.slice/crio-6796fd1468ed5765d1591619c6977ac43e473ce39f59c2060d221f57ca3f0e4b WatchSource:0}: Error finding container 6796fd1468ed5765d1591619c6977ac43e473ce39f59c2060d221f57ca3f0e4b: Status 404 returned error can't find the container with id 6796fd1468ed5765d1591619c6977ac43e473ce39f59c2060d221f57ca3f0e4b Dec 03 00:43:42 crc kubenswrapper[4953]: I1203 00:43:42.126994 4953 generic.go:334] "Generic (PLEG): container finished" podID="8f4dbfe6-01d5-43dc-9f8b-a1f78aa2397b" containerID="e7997201bee6ddbb0a7e9cea5829e87511232ab84b93480dbdbaed0f6a8d3418" exitCode=0 Dec 03 00:43:42 crc kubenswrapper[4953]: I1203 00:43:42.127094 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jtckq" event={"ID":"8f4dbfe6-01d5-43dc-9f8b-a1f78aa2397b","Type":"ContainerDied","Data":"e7997201bee6ddbb0a7e9cea5829e87511232ab84b93480dbdbaed0f6a8d3418"} Dec 03 00:43:42 crc kubenswrapper[4953]: I1203 00:43:42.127172 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jtckq" event={"ID":"8f4dbfe6-01d5-43dc-9f8b-a1f78aa2397b","Type":"ContainerStarted","Data":"6796fd1468ed5765d1591619c6977ac43e473ce39f59c2060d221f57ca3f0e4b"} Dec 03 00:43:43 crc kubenswrapper[4953]: I1203 00:43:43.140881 4953 generic.go:334] "Generic (PLEG): container finished" podID="3bb49b8a-e497-48bf-bc83-5effc81b78e4" containerID="50f5cfa6b05f2d2ca54ad7d5d26191e72f9584044d74a4af470fd241779d348b" exitCode=0 Dec 03 00:43:43 crc kubenswrapper[4953]: I1203 00:43:43.141115 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-r4xg5" event={"ID":"3bb49b8a-e497-48bf-bc83-5effc81b78e4","Type":"ContainerDied","Data":"50f5cfa6b05f2d2ca54ad7d5d26191e72f9584044d74a4af470fd241779d348b"} Dec 03 00:43:43 crc kubenswrapper[4953]: I1203 00:43:43.146166 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jtckq" event={"ID":"8f4dbfe6-01d5-43dc-9f8b-a1f78aa2397b","Type":"ContainerStarted","Data":"3e526e4b730d9867f57876c99e4a77b3c67e280bb3fcbcc3dbd4ce727d819fec"} Dec 03 00:43:44 crc kubenswrapper[4953]: I1203 00:43:44.157851 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-r4xg5" event={"ID":"3bb49b8a-e497-48bf-bc83-5effc81b78e4","Type":"ContainerStarted","Data":"7085cddfa158d9c7d3a6dd7f83194af47b72439b183b123905b5f6cde455e8a8"} Dec 03 00:43:44 crc kubenswrapper[4953]: I1203 00:43:44.160335 4953 generic.go:334] "Generic (PLEG): container finished" podID="8f4dbfe6-01d5-43dc-9f8b-a1f78aa2397b" containerID="3e526e4b730d9867f57876c99e4a77b3c67e280bb3fcbcc3dbd4ce727d819fec" exitCode=0 Dec 03 00:43:44 crc kubenswrapper[4953]: I1203 00:43:44.160392 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jtckq" event={"ID":"8f4dbfe6-01d5-43dc-9f8b-a1f78aa2397b","Type":"ContainerDied","Data":"3e526e4b730d9867f57876c99e4a77b3c67e280bb3fcbcc3dbd4ce727d819fec"} Dec 03 00:43:44 crc kubenswrapper[4953]: I1203 00:43:44.182617 4953 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-r4xg5" podStartSLOduration=2.637975493 podStartE2EDuration="6.182592509s" podCreationTimestamp="2025-12-03 00:43:38 +0000 UTC" firstStartedPulling="2025-12-03 00:43:40.029596335 +0000 UTC m=+2209.013877118" lastFinishedPulling="2025-12-03 00:43:43.574213351 +0000 UTC m=+2212.558494134" observedRunningTime="2025-12-03 00:43:44.180253321 +0000 UTC m=+2213.164534104" watchObservedRunningTime="2025-12-03 00:43:44.182592509 +0000 UTC m=+2213.166873292" Dec 03 00:43:45 crc kubenswrapper[4953]: I1203 00:43:45.174955 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jtckq" event={"ID":"8f4dbfe6-01d5-43dc-9f8b-a1f78aa2397b","Type":"ContainerStarted","Data":"cd61ee824aed2844b53f604cf6e52168b7987bf5882f4b6a77d0ea81b3c10ebe"} Dec 03 00:43:45 crc kubenswrapper[4953]: I1203 00:43:45.212747 4953 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-jtckq" podStartSLOduration=2.710848241 podStartE2EDuration="5.212712757s" podCreationTimestamp="2025-12-03 00:43:40 +0000 UTC" firstStartedPulling="2025-12-03 00:43:42.130059435 +0000 UTC m=+2211.114340218" lastFinishedPulling="2025-12-03 00:43:44.631923951 +0000 UTC m=+2213.616204734" observedRunningTime="2025-12-03 00:43:45.206591656 +0000 UTC m=+2214.190872429" watchObservedRunningTime="2025-12-03 00:43:45.212712757 +0000 UTC m=+2214.196993540" Dec 03 00:43:49 crc kubenswrapper[4953]: I1203 00:43:49.105711 4953 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-r4xg5" Dec 03 00:43:49 crc kubenswrapper[4953]: I1203 00:43:49.106276 4953 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-r4xg5" Dec 03 00:43:50 crc kubenswrapper[4953]: I1203 00:43:50.154853 4953 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-r4xg5" podUID="3bb49b8a-e497-48bf-bc83-5effc81b78e4" containerName="registry-server" probeResult="failure" output=< Dec 03 00:43:50 crc kubenswrapper[4953]: timeout: failed to connect service ":50051" within 1s Dec 03 00:43:50 crc kubenswrapper[4953]: > Dec 03 00:43:50 crc kubenswrapper[4953]: I1203 00:43:50.873276 4953 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-jtckq" Dec 03 00:43:50 crc kubenswrapper[4953]: I1203 00:43:50.873355 4953 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-jtckq" Dec 03 00:43:50 crc kubenswrapper[4953]: I1203 00:43:50.921103 4953 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-jtckq" Dec 03 00:43:51 crc kubenswrapper[4953]: I1203 00:43:51.277037 4953 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-jtckq" Dec 03 00:43:51 crc kubenswrapper[4953]: I1203 00:43:51.341299 4953 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-jtckq"] Dec 03 00:43:53 crc kubenswrapper[4953]: I1203 00:43:53.243413 4953 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-jtckq" podUID="8f4dbfe6-01d5-43dc-9f8b-a1f78aa2397b" containerName="registry-server" containerID="cri-o://cd61ee824aed2844b53f604cf6e52168b7987bf5882f4b6a77d0ea81b3c10ebe" gracePeriod=2 Dec 03 00:43:54 crc kubenswrapper[4953]: E1203 00:43:54.883300 4953 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7ee12f05_7ff1_4efa_b45e_45376c77b906.slice/crio-conmon-a778f07eebaf40f3618ceb12ade247a3b45a9bfabd835e0b96f52a8325be6ba2.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7ee12f05_7ff1_4efa_b45e_45376c77b906.slice/crio-a778f07eebaf40f3618ceb12ade247a3b45a9bfabd835e0b96f52a8325be6ba2.scope\": RecentStats: unable to find data in memory cache]" Dec 03 00:43:55 crc kubenswrapper[4953]: I1203 00:43:55.261516 4953 generic.go:334] "Generic (PLEG): container finished" podID="8f4dbfe6-01d5-43dc-9f8b-a1f78aa2397b" containerID="cd61ee824aed2844b53f604cf6e52168b7987bf5882f4b6a77d0ea81b3c10ebe" exitCode=0 Dec 03 00:43:55 crc kubenswrapper[4953]: I1203 00:43:55.261572 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jtckq" event={"ID":"8f4dbfe6-01d5-43dc-9f8b-a1f78aa2397b","Type":"ContainerDied","Data":"cd61ee824aed2844b53f604cf6e52168b7987bf5882f4b6a77d0ea81b3c10ebe"} Dec 03 00:43:55 crc kubenswrapper[4953]: I1203 00:43:55.264290 4953 generic.go:334] "Generic (PLEG): container finished" podID="7ee12f05-7ff1-4efa-b45e-45376c77b906" containerID="a778f07eebaf40f3618ceb12ade247a3b45a9bfabd835e0b96f52a8325be6ba2" exitCode=0 Dec 03 00:43:55 crc kubenswrapper[4953]: I1203 00:43:55.264368 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-mnrv5/must-gather-s69cf" event={"ID":"7ee12f05-7ff1-4efa-b45e-45376c77b906","Type":"ContainerDied","Data":"a778f07eebaf40f3618ceb12ade247a3b45a9bfabd835e0b96f52a8325be6ba2"} Dec 03 00:43:55 crc kubenswrapper[4953]: I1203 00:43:55.265221 4953 scope.go:117] "RemoveContainer" containerID="a778f07eebaf40f3618ceb12ade247a3b45a9bfabd835e0b96f52a8325be6ba2" Dec 03 00:43:55 crc kubenswrapper[4953]: I1203 00:43:55.478650 4953 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-jtckq" Dec 03 00:43:55 crc kubenswrapper[4953]: I1203 00:43:55.646591 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8f4dbfe6-01d5-43dc-9f8b-a1f78aa2397b-utilities\") pod \"8f4dbfe6-01d5-43dc-9f8b-a1f78aa2397b\" (UID: \"8f4dbfe6-01d5-43dc-9f8b-a1f78aa2397b\") " Dec 03 00:43:55 crc kubenswrapper[4953]: I1203 00:43:55.646898 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wzgdj\" (UniqueName: \"kubernetes.io/projected/8f4dbfe6-01d5-43dc-9f8b-a1f78aa2397b-kube-api-access-wzgdj\") pod \"8f4dbfe6-01d5-43dc-9f8b-a1f78aa2397b\" (UID: \"8f4dbfe6-01d5-43dc-9f8b-a1f78aa2397b\") " Dec 03 00:43:55 crc kubenswrapper[4953]: I1203 00:43:55.647040 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8f4dbfe6-01d5-43dc-9f8b-a1f78aa2397b-catalog-content\") pod \"8f4dbfe6-01d5-43dc-9f8b-a1f78aa2397b\" (UID: \"8f4dbfe6-01d5-43dc-9f8b-a1f78aa2397b\") " Dec 03 00:43:55 crc kubenswrapper[4953]: I1203 00:43:55.647672 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f4dbfe6-01d5-43dc-9f8b-a1f78aa2397b-utilities" (OuterVolumeSpecName: "utilities") pod "8f4dbfe6-01d5-43dc-9f8b-a1f78aa2397b" (UID: "8f4dbfe6-01d5-43dc-9f8b-a1f78aa2397b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 00:43:55 crc kubenswrapper[4953]: I1203 00:43:55.656299 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f4dbfe6-01d5-43dc-9f8b-a1f78aa2397b-kube-api-access-wzgdj" (OuterVolumeSpecName: "kube-api-access-wzgdj") pod "8f4dbfe6-01d5-43dc-9f8b-a1f78aa2397b" (UID: "8f4dbfe6-01d5-43dc-9f8b-a1f78aa2397b"). InnerVolumeSpecName "kube-api-access-wzgdj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:43:55 crc kubenswrapper[4953]: I1203 00:43:55.706154 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f4dbfe6-01d5-43dc-9f8b-a1f78aa2397b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8f4dbfe6-01d5-43dc-9f8b-a1f78aa2397b" (UID: "8f4dbfe6-01d5-43dc-9f8b-a1f78aa2397b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 00:43:55 crc kubenswrapper[4953]: I1203 00:43:55.749534 4953 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wzgdj\" (UniqueName: \"kubernetes.io/projected/8f4dbfe6-01d5-43dc-9f8b-a1f78aa2397b-kube-api-access-wzgdj\") on node \"crc\" DevicePath \"\"" Dec 03 00:43:55 crc kubenswrapper[4953]: I1203 00:43:55.749653 4953 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8f4dbfe6-01d5-43dc-9f8b-a1f78aa2397b-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 00:43:55 crc kubenswrapper[4953]: I1203 00:43:55.749699 4953 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8f4dbfe6-01d5-43dc-9f8b-a1f78aa2397b-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 00:43:55 crc kubenswrapper[4953]: I1203 00:43:55.844304 4953 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-mnrv5_must-gather-s69cf_7ee12f05-7ff1-4efa-b45e-45376c77b906/gather/0.log" Dec 03 00:43:56 crc kubenswrapper[4953]: I1203 00:43:56.275470 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jtckq" event={"ID":"8f4dbfe6-01d5-43dc-9f8b-a1f78aa2397b","Type":"ContainerDied","Data":"6796fd1468ed5765d1591619c6977ac43e473ce39f59c2060d221f57ca3f0e4b"} Dec 03 00:43:56 crc kubenswrapper[4953]: I1203 00:43:56.275561 4953 scope.go:117] "RemoveContainer" containerID="cd61ee824aed2844b53f604cf6e52168b7987bf5882f4b6a77d0ea81b3c10ebe" Dec 03 00:43:56 crc kubenswrapper[4953]: I1203 00:43:56.275612 4953 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-jtckq" Dec 03 00:43:56 crc kubenswrapper[4953]: I1203 00:43:56.298569 4953 scope.go:117] "RemoveContainer" containerID="3e526e4b730d9867f57876c99e4a77b3c67e280bb3fcbcc3dbd4ce727d819fec" Dec 03 00:43:56 crc kubenswrapper[4953]: I1203 00:43:56.318121 4953 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-jtckq"] Dec 03 00:43:56 crc kubenswrapper[4953]: I1203 00:43:56.328202 4953 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-jtckq"] Dec 03 00:43:56 crc kubenswrapper[4953]: I1203 00:43:56.339790 4953 scope.go:117] "RemoveContainer" containerID="e7997201bee6ddbb0a7e9cea5829e87511232ab84b93480dbdbaed0f6a8d3418" Dec 03 00:43:57 crc kubenswrapper[4953]: I1203 00:43:57.357304 4953 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f4dbfe6-01d5-43dc-9f8b-a1f78aa2397b" path="/var/lib/kubelet/pods/8f4dbfe6-01d5-43dc-9f8b-a1f78aa2397b/volumes" Dec 03 00:43:59 crc kubenswrapper[4953]: I1203 00:43:59.163317 4953 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-r4xg5" Dec 03 00:43:59 crc kubenswrapper[4953]: I1203 00:43:59.213850 4953 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-r4xg5" Dec 03 00:43:59 crc kubenswrapper[4953]: I1203 00:43:59.723863 4953 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-r4xg5"] Dec 03 00:44:00 crc kubenswrapper[4953]: I1203 00:44:00.314073 4953 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-r4xg5" podUID="3bb49b8a-e497-48bf-bc83-5effc81b78e4" containerName="registry-server" containerID="cri-o://7085cddfa158d9c7d3a6dd7f83194af47b72439b183b123905b5f6cde455e8a8" gracePeriod=2 Dec 03 00:44:00 crc kubenswrapper[4953]: I1203 00:44:00.771211 4953 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-r4xg5" Dec 03 00:44:00 crc kubenswrapper[4953]: I1203 00:44:00.898523 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3bb49b8a-e497-48bf-bc83-5effc81b78e4-catalog-content\") pod \"3bb49b8a-e497-48bf-bc83-5effc81b78e4\" (UID: \"3bb49b8a-e497-48bf-bc83-5effc81b78e4\") " Dec 03 00:44:00 crc kubenswrapper[4953]: I1203 00:44:00.899198 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3bb49b8a-e497-48bf-bc83-5effc81b78e4-utilities\") pod \"3bb49b8a-e497-48bf-bc83-5effc81b78e4\" (UID: \"3bb49b8a-e497-48bf-bc83-5effc81b78e4\") " Dec 03 00:44:00 crc kubenswrapper[4953]: I1203 00:44:00.899280 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fmxmz\" (UniqueName: \"kubernetes.io/projected/3bb49b8a-e497-48bf-bc83-5effc81b78e4-kube-api-access-fmxmz\") pod \"3bb49b8a-e497-48bf-bc83-5effc81b78e4\" (UID: \"3bb49b8a-e497-48bf-bc83-5effc81b78e4\") " Dec 03 00:44:00 crc kubenswrapper[4953]: I1203 00:44:00.900353 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3bb49b8a-e497-48bf-bc83-5effc81b78e4-utilities" (OuterVolumeSpecName: "utilities") pod "3bb49b8a-e497-48bf-bc83-5effc81b78e4" (UID: "3bb49b8a-e497-48bf-bc83-5effc81b78e4"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 00:44:00 crc kubenswrapper[4953]: I1203 00:44:00.920802 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3bb49b8a-e497-48bf-bc83-5effc81b78e4-kube-api-access-fmxmz" (OuterVolumeSpecName: "kube-api-access-fmxmz") pod "3bb49b8a-e497-48bf-bc83-5effc81b78e4" (UID: "3bb49b8a-e497-48bf-bc83-5effc81b78e4"). InnerVolumeSpecName "kube-api-access-fmxmz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:44:01 crc kubenswrapper[4953]: I1203 00:44:01.002228 4953 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3bb49b8a-e497-48bf-bc83-5effc81b78e4-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 00:44:01 crc kubenswrapper[4953]: I1203 00:44:01.002290 4953 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fmxmz\" (UniqueName: \"kubernetes.io/projected/3bb49b8a-e497-48bf-bc83-5effc81b78e4-kube-api-access-fmxmz\") on node \"crc\" DevicePath \"\"" Dec 03 00:44:01 crc kubenswrapper[4953]: I1203 00:44:01.023583 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3bb49b8a-e497-48bf-bc83-5effc81b78e4-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "3bb49b8a-e497-48bf-bc83-5effc81b78e4" (UID: "3bb49b8a-e497-48bf-bc83-5effc81b78e4"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 00:44:01 crc kubenswrapper[4953]: I1203 00:44:01.103486 4953 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3bb49b8a-e497-48bf-bc83-5effc81b78e4-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 00:44:01 crc kubenswrapper[4953]: I1203 00:44:01.326506 4953 generic.go:334] "Generic (PLEG): container finished" podID="3bb49b8a-e497-48bf-bc83-5effc81b78e4" containerID="7085cddfa158d9c7d3a6dd7f83194af47b72439b183b123905b5f6cde455e8a8" exitCode=0 Dec 03 00:44:01 crc kubenswrapper[4953]: I1203 00:44:01.326566 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-r4xg5" event={"ID":"3bb49b8a-e497-48bf-bc83-5effc81b78e4","Type":"ContainerDied","Data":"7085cddfa158d9c7d3a6dd7f83194af47b72439b183b123905b5f6cde455e8a8"} Dec 03 00:44:01 crc kubenswrapper[4953]: I1203 00:44:01.326588 4953 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-r4xg5" Dec 03 00:44:01 crc kubenswrapper[4953]: I1203 00:44:01.326601 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-r4xg5" event={"ID":"3bb49b8a-e497-48bf-bc83-5effc81b78e4","Type":"ContainerDied","Data":"1840c550eaa1c549084678e0b215d09bd1589b9282eb2e544c0224cee3f46434"} Dec 03 00:44:01 crc kubenswrapper[4953]: I1203 00:44:01.326623 4953 scope.go:117] "RemoveContainer" containerID="7085cddfa158d9c7d3a6dd7f83194af47b72439b183b123905b5f6cde455e8a8" Dec 03 00:44:01 crc kubenswrapper[4953]: I1203 00:44:01.351266 4953 scope.go:117] "RemoveContainer" containerID="50f5cfa6b05f2d2ca54ad7d5d26191e72f9584044d74a4af470fd241779d348b" Dec 03 00:44:01 crc kubenswrapper[4953]: I1203 00:44:01.413066 4953 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-r4xg5"] Dec 03 00:44:01 crc kubenswrapper[4953]: I1203 00:44:01.413764 4953 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-r4xg5"] Dec 03 00:44:01 crc kubenswrapper[4953]: I1203 00:44:01.418237 4953 scope.go:117] "RemoveContainer" containerID="485392e1c2b391f689c44dc861da19005b55bcb9d5c594ba3c31a2eed7897ad1" Dec 03 00:44:01 crc kubenswrapper[4953]: I1203 00:44:01.443492 4953 scope.go:117] "RemoveContainer" containerID="7085cddfa158d9c7d3a6dd7f83194af47b72439b183b123905b5f6cde455e8a8" Dec 03 00:44:01 crc kubenswrapper[4953]: E1203 00:44:01.444280 4953 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7085cddfa158d9c7d3a6dd7f83194af47b72439b183b123905b5f6cde455e8a8\": container with ID starting with 7085cddfa158d9c7d3a6dd7f83194af47b72439b183b123905b5f6cde455e8a8 not found: ID does not exist" containerID="7085cddfa158d9c7d3a6dd7f83194af47b72439b183b123905b5f6cde455e8a8" Dec 03 00:44:01 crc kubenswrapper[4953]: I1203 00:44:01.444388 4953 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7085cddfa158d9c7d3a6dd7f83194af47b72439b183b123905b5f6cde455e8a8"} err="failed to get container status \"7085cddfa158d9c7d3a6dd7f83194af47b72439b183b123905b5f6cde455e8a8\": rpc error: code = NotFound desc = could not find container \"7085cddfa158d9c7d3a6dd7f83194af47b72439b183b123905b5f6cde455e8a8\": container with ID starting with 7085cddfa158d9c7d3a6dd7f83194af47b72439b183b123905b5f6cde455e8a8 not found: ID does not exist" Dec 03 00:44:01 crc kubenswrapper[4953]: I1203 00:44:01.444563 4953 scope.go:117] "RemoveContainer" containerID="50f5cfa6b05f2d2ca54ad7d5d26191e72f9584044d74a4af470fd241779d348b" Dec 03 00:44:01 crc kubenswrapper[4953]: E1203 00:44:01.445178 4953 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"50f5cfa6b05f2d2ca54ad7d5d26191e72f9584044d74a4af470fd241779d348b\": container with ID starting with 50f5cfa6b05f2d2ca54ad7d5d26191e72f9584044d74a4af470fd241779d348b not found: ID does not exist" containerID="50f5cfa6b05f2d2ca54ad7d5d26191e72f9584044d74a4af470fd241779d348b" Dec 03 00:44:01 crc kubenswrapper[4953]: I1203 00:44:01.445295 4953 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"50f5cfa6b05f2d2ca54ad7d5d26191e72f9584044d74a4af470fd241779d348b"} err="failed to get container status \"50f5cfa6b05f2d2ca54ad7d5d26191e72f9584044d74a4af470fd241779d348b\": rpc error: code = NotFound desc = could not find container \"50f5cfa6b05f2d2ca54ad7d5d26191e72f9584044d74a4af470fd241779d348b\": container with ID starting with 50f5cfa6b05f2d2ca54ad7d5d26191e72f9584044d74a4af470fd241779d348b not found: ID does not exist" Dec 03 00:44:01 crc kubenswrapper[4953]: I1203 00:44:01.445376 4953 scope.go:117] "RemoveContainer" containerID="485392e1c2b391f689c44dc861da19005b55bcb9d5c594ba3c31a2eed7897ad1" Dec 03 00:44:01 crc kubenswrapper[4953]: E1203 00:44:01.445706 4953 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"485392e1c2b391f689c44dc861da19005b55bcb9d5c594ba3c31a2eed7897ad1\": container with ID starting with 485392e1c2b391f689c44dc861da19005b55bcb9d5c594ba3c31a2eed7897ad1 not found: ID does not exist" containerID="485392e1c2b391f689c44dc861da19005b55bcb9d5c594ba3c31a2eed7897ad1" Dec 03 00:44:01 crc kubenswrapper[4953]: I1203 00:44:01.445806 4953 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"485392e1c2b391f689c44dc861da19005b55bcb9d5c594ba3c31a2eed7897ad1"} err="failed to get container status \"485392e1c2b391f689c44dc861da19005b55bcb9d5c594ba3c31a2eed7897ad1\": rpc error: code = NotFound desc = could not find container \"485392e1c2b391f689c44dc861da19005b55bcb9d5c594ba3c31a2eed7897ad1\": container with ID starting with 485392e1c2b391f689c44dc861da19005b55bcb9d5c594ba3c31a2eed7897ad1 not found: ID does not exist" Dec 03 00:44:02 crc kubenswrapper[4953]: I1203 00:44:02.540714 4953 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-mnrv5/must-gather-s69cf"] Dec 03 00:44:02 crc kubenswrapper[4953]: I1203 00:44:02.541215 4953 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-mnrv5/must-gather-s69cf" podUID="7ee12f05-7ff1-4efa-b45e-45376c77b906" containerName="copy" containerID="cri-o://02447eb8bbe00c7b77ea6b8275874ea50738776b27034961751231aebfe8efc1" gracePeriod=2 Dec 03 00:44:02 crc kubenswrapper[4953]: I1203 00:44:02.549197 4953 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-mnrv5/must-gather-s69cf"] Dec 03 00:44:02 crc kubenswrapper[4953]: I1203 00:44:02.957431 4953 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-mnrv5_must-gather-s69cf_7ee12f05-7ff1-4efa-b45e-45376c77b906/copy/0.log" Dec 03 00:44:02 crc kubenswrapper[4953]: I1203 00:44:02.958485 4953 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-mnrv5/must-gather-s69cf" Dec 03 00:44:02 crc kubenswrapper[4953]: I1203 00:44:02.983228 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/7ee12f05-7ff1-4efa-b45e-45376c77b906-must-gather-output\") pod \"7ee12f05-7ff1-4efa-b45e-45376c77b906\" (UID: \"7ee12f05-7ff1-4efa-b45e-45376c77b906\") " Dec 03 00:44:02 crc kubenswrapper[4953]: I1203 00:44:02.994634 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jtc9v\" (UniqueName: \"kubernetes.io/projected/7ee12f05-7ff1-4efa-b45e-45376c77b906-kube-api-access-jtc9v\") pod \"7ee12f05-7ff1-4efa-b45e-45376c77b906\" (UID: \"7ee12f05-7ff1-4efa-b45e-45376c77b906\") " Dec 03 00:44:03 crc kubenswrapper[4953]: I1203 00:44:03.003174 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7ee12f05-7ff1-4efa-b45e-45376c77b906-kube-api-access-jtc9v" (OuterVolumeSpecName: "kube-api-access-jtc9v") pod "7ee12f05-7ff1-4efa-b45e-45376c77b906" (UID: "7ee12f05-7ff1-4efa-b45e-45376c77b906"). InnerVolumeSpecName "kube-api-access-jtc9v". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:44:03 crc kubenswrapper[4953]: I1203 00:44:03.057148 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7ee12f05-7ff1-4efa-b45e-45376c77b906-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "7ee12f05-7ff1-4efa-b45e-45376c77b906" (UID: "7ee12f05-7ff1-4efa-b45e-45376c77b906"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 00:44:03 crc kubenswrapper[4953]: I1203 00:44:03.098155 4953 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jtc9v\" (UniqueName: \"kubernetes.io/projected/7ee12f05-7ff1-4efa-b45e-45376c77b906-kube-api-access-jtc9v\") on node \"crc\" DevicePath \"\"" Dec 03 00:44:03 crc kubenswrapper[4953]: I1203 00:44:03.098194 4953 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/7ee12f05-7ff1-4efa-b45e-45376c77b906-must-gather-output\") on node \"crc\" DevicePath \"\"" Dec 03 00:44:03 crc kubenswrapper[4953]: I1203 00:44:03.397322 4953 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-mnrv5_must-gather-s69cf_7ee12f05-7ff1-4efa-b45e-45376c77b906/copy/0.log" Dec 03 00:44:03 crc kubenswrapper[4953]: I1203 00:44:03.398673 4953 generic.go:334] "Generic (PLEG): container finished" podID="7ee12f05-7ff1-4efa-b45e-45376c77b906" containerID="02447eb8bbe00c7b77ea6b8275874ea50738776b27034961751231aebfe8efc1" exitCode=143 Dec 03 00:44:03 crc kubenswrapper[4953]: I1203 00:44:03.399946 4953 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-mnrv5/must-gather-s69cf" Dec 03 00:44:03 crc kubenswrapper[4953]: I1203 00:44:03.405403 4953 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3bb49b8a-e497-48bf-bc83-5effc81b78e4" path="/var/lib/kubelet/pods/3bb49b8a-e497-48bf-bc83-5effc81b78e4/volumes" Dec 03 00:44:03 crc kubenswrapper[4953]: I1203 00:44:03.407067 4953 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7ee12f05-7ff1-4efa-b45e-45376c77b906" path="/var/lib/kubelet/pods/7ee12f05-7ff1-4efa-b45e-45376c77b906/volumes" Dec 03 00:44:03 crc kubenswrapper[4953]: I1203 00:44:03.407926 4953 scope.go:117] "RemoveContainer" containerID="02447eb8bbe00c7b77ea6b8275874ea50738776b27034961751231aebfe8efc1" Dec 03 00:44:03 crc kubenswrapper[4953]: I1203 00:44:03.446429 4953 scope.go:117] "RemoveContainer" containerID="a778f07eebaf40f3618ceb12ade247a3b45a9bfabd835e0b96f52a8325be6ba2" Dec 03 00:44:03 crc kubenswrapper[4953]: I1203 00:44:03.500948 4953 scope.go:117] "RemoveContainer" containerID="02447eb8bbe00c7b77ea6b8275874ea50738776b27034961751231aebfe8efc1" Dec 03 00:44:03 crc kubenswrapper[4953]: E1203 00:44:03.501719 4953 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"02447eb8bbe00c7b77ea6b8275874ea50738776b27034961751231aebfe8efc1\": container with ID starting with 02447eb8bbe00c7b77ea6b8275874ea50738776b27034961751231aebfe8efc1 not found: ID does not exist" containerID="02447eb8bbe00c7b77ea6b8275874ea50738776b27034961751231aebfe8efc1" Dec 03 00:44:03 crc kubenswrapper[4953]: I1203 00:44:03.501774 4953 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"02447eb8bbe00c7b77ea6b8275874ea50738776b27034961751231aebfe8efc1"} err="failed to get container status \"02447eb8bbe00c7b77ea6b8275874ea50738776b27034961751231aebfe8efc1\": rpc error: code = NotFound desc = could not find container \"02447eb8bbe00c7b77ea6b8275874ea50738776b27034961751231aebfe8efc1\": container with ID starting with 02447eb8bbe00c7b77ea6b8275874ea50738776b27034961751231aebfe8efc1 not found: ID does not exist" Dec 03 00:44:03 crc kubenswrapper[4953]: I1203 00:44:03.501811 4953 scope.go:117] "RemoveContainer" containerID="a778f07eebaf40f3618ceb12ade247a3b45a9bfabd835e0b96f52a8325be6ba2" Dec 03 00:44:03 crc kubenswrapper[4953]: E1203 00:44:03.502281 4953 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a778f07eebaf40f3618ceb12ade247a3b45a9bfabd835e0b96f52a8325be6ba2\": container with ID starting with a778f07eebaf40f3618ceb12ade247a3b45a9bfabd835e0b96f52a8325be6ba2 not found: ID does not exist" containerID="a778f07eebaf40f3618ceb12ade247a3b45a9bfabd835e0b96f52a8325be6ba2" Dec 03 00:44:03 crc kubenswrapper[4953]: I1203 00:44:03.502382 4953 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a778f07eebaf40f3618ceb12ade247a3b45a9bfabd835e0b96f52a8325be6ba2"} err="failed to get container status \"a778f07eebaf40f3618ceb12ade247a3b45a9bfabd835e0b96f52a8325be6ba2\": rpc error: code = NotFound desc = could not find container \"a778f07eebaf40f3618ceb12ade247a3b45a9bfabd835e0b96f52a8325be6ba2\": container with ID starting with a778f07eebaf40f3618ceb12ade247a3b45a9bfabd835e0b96f52a8325be6ba2 not found: ID does not exist" Dec 03 00:45:00 crc kubenswrapper[4953]: I1203 00:45:00.149729 4953 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412045-b8q4q"] Dec 03 00:45:00 crc kubenswrapper[4953]: E1203 00:45:00.150749 4953 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8f4dbfe6-01d5-43dc-9f8b-a1f78aa2397b" containerName="extract-utilities" Dec 03 00:45:00 crc kubenswrapper[4953]: I1203 00:45:00.150765 4953 state_mem.go:107] "Deleted CPUSet assignment" podUID="8f4dbfe6-01d5-43dc-9f8b-a1f78aa2397b" containerName="extract-utilities" Dec 03 00:45:00 crc kubenswrapper[4953]: E1203 00:45:00.150782 4953 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3bb49b8a-e497-48bf-bc83-5effc81b78e4" containerName="registry-server" Dec 03 00:45:00 crc kubenswrapper[4953]: I1203 00:45:00.150789 4953 state_mem.go:107] "Deleted CPUSet assignment" podUID="3bb49b8a-e497-48bf-bc83-5effc81b78e4" containerName="registry-server" Dec 03 00:45:00 crc kubenswrapper[4953]: E1203 00:45:00.150805 4953 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8f4dbfe6-01d5-43dc-9f8b-a1f78aa2397b" containerName="extract-content" Dec 03 00:45:00 crc kubenswrapper[4953]: I1203 00:45:00.150813 4953 state_mem.go:107] "Deleted CPUSet assignment" podUID="8f4dbfe6-01d5-43dc-9f8b-a1f78aa2397b" containerName="extract-content" Dec 03 00:45:00 crc kubenswrapper[4953]: E1203 00:45:00.150825 4953 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7ee12f05-7ff1-4efa-b45e-45376c77b906" containerName="gather" Dec 03 00:45:00 crc kubenswrapper[4953]: I1203 00:45:00.150830 4953 state_mem.go:107] "Deleted CPUSet assignment" podUID="7ee12f05-7ff1-4efa-b45e-45376c77b906" containerName="gather" Dec 03 00:45:00 crc kubenswrapper[4953]: E1203 00:45:00.150839 4953 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7ee12f05-7ff1-4efa-b45e-45376c77b906" containerName="copy" Dec 03 00:45:00 crc kubenswrapper[4953]: I1203 00:45:00.150845 4953 state_mem.go:107] "Deleted CPUSet assignment" podUID="7ee12f05-7ff1-4efa-b45e-45376c77b906" containerName="copy" Dec 03 00:45:00 crc kubenswrapper[4953]: E1203 00:45:00.150852 4953 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8f4dbfe6-01d5-43dc-9f8b-a1f78aa2397b" containerName="registry-server" Dec 03 00:45:00 crc kubenswrapper[4953]: I1203 00:45:00.150859 4953 state_mem.go:107] "Deleted CPUSet assignment" podUID="8f4dbfe6-01d5-43dc-9f8b-a1f78aa2397b" containerName="registry-server" Dec 03 00:45:00 crc kubenswrapper[4953]: E1203 00:45:00.150868 4953 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3bb49b8a-e497-48bf-bc83-5effc81b78e4" containerName="extract-utilities" Dec 03 00:45:00 crc kubenswrapper[4953]: I1203 00:45:00.150876 4953 state_mem.go:107] "Deleted CPUSet assignment" podUID="3bb49b8a-e497-48bf-bc83-5effc81b78e4" containerName="extract-utilities" Dec 03 00:45:00 crc kubenswrapper[4953]: E1203 00:45:00.150886 4953 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3bb49b8a-e497-48bf-bc83-5effc81b78e4" containerName="extract-content" Dec 03 00:45:00 crc kubenswrapper[4953]: I1203 00:45:00.150892 4953 state_mem.go:107] "Deleted CPUSet assignment" podUID="3bb49b8a-e497-48bf-bc83-5effc81b78e4" containerName="extract-content" Dec 03 00:45:00 crc kubenswrapper[4953]: I1203 00:45:00.151027 4953 memory_manager.go:354] "RemoveStaleState removing state" podUID="3bb49b8a-e497-48bf-bc83-5effc81b78e4" containerName="registry-server" Dec 03 00:45:00 crc kubenswrapper[4953]: I1203 00:45:00.151040 4953 memory_manager.go:354] "RemoveStaleState removing state" podUID="8f4dbfe6-01d5-43dc-9f8b-a1f78aa2397b" containerName="registry-server" Dec 03 00:45:00 crc kubenswrapper[4953]: I1203 00:45:00.151058 4953 memory_manager.go:354] "RemoveStaleState removing state" podUID="7ee12f05-7ff1-4efa-b45e-45376c77b906" containerName="copy" Dec 03 00:45:00 crc kubenswrapper[4953]: I1203 00:45:00.151069 4953 memory_manager.go:354] "RemoveStaleState removing state" podUID="7ee12f05-7ff1-4efa-b45e-45376c77b906" containerName="gather" Dec 03 00:45:00 crc kubenswrapper[4953]: I1203 00:45:00.151515 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29412045-b8q4q" Dec 03 00:45:00 crc kubenswrapper[4953]: I1203 00:45:00.155120 4953 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 03 00:45:00 crc kubenswrapper[4953]: I1203 00:45:00.155484 4953 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 03 00:45:00 crc kubenswrapper[4953]: I1203 00:45:00.171759 4953 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412045-b8q4q"] Dec 03 00:45:00 crc kubenswrapper[4953]: I1203 00:45:00.233158 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/df482960-425e-4b6b-9284-cedcc918cc94-config-volume\") pod \"collect-profiles-29412045-b8q4q\" (UID: \"df482960-425e-4b6b-9284-cedcc918cc94\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412045-b8q4q" Dec 03 00:45:00 crc kubenswrapper[4953]: I1203 00:45:00.233204 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5sj8c\" (UniqueName: \"kubernetes.io/projected/df482960-425e-4b6b-9284-cedcc918cc94-kube-api-access-5sj8c\") pod \"collect-profiles-29412045-b8q4q\" (UID: \"df482960-425e-4b6b-9284-cedcc918cc94\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412045-b8q4q" Dec 03 00:45:00 crc kubenswrapper[4953]: I1203 00:45:00.233245 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/df482960-425e-4b6b-9284-cedcc918cc94-secret-volume\") pod \"collect-profiles-29412045-b8q4q\" (UID: \"df482960-425e-4b6b-9284-cedcc918cc94\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412045-b8q4q" Dec 03 00:45:00 crc kubenswrapper[4953]: I1203 00:45:00.335235 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/df482960-425e-4b6b-9284-cedcc918cc94-secret-volume\") pod \"collect-profiles-29412045-b8q4q\" (UID: \"df482960-425e-4b6b-9284-cedcc918cc94\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412045-b8q4q" Dec 03 00:45:00 crc kubenswrapper[4953]: I1203 00:45:00.335371 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/df482960-425e-4b6b-9284-cedcc918cc94-config-volume\") pod \"collect-profiles-29412045-b8q4q\" (UID: \"df482960-425e-4b6b-9284-cedcc918cc94\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412045-b8q4q" Dec 03 00:45:00 crc kubenswrapper[4953]: I1203 00:45:00.335400 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5sj8c\" (UniqueName: \"kubernetes.io/projected/df482960-425e-4b6b-9284-cedcc918cc94-kube-api-access-5sj8c\") pod \"collect-profiles-29412045-b8q4q\" (UID: \"df482960-425e-4b6b-9284-cedcc918cc94\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412045-b8q4q" Dec 03 00:45:00 crc kubenswrapper[4953]: I1203 00:45:00.336851 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/df482960-425e-4b6b-9284-cedcc918cc94-config-volume\") pod \"collect-profiles-29412045-b8q4q\" (UID: \"df482960-425e-4b6b-9284-cedcc918cc94\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412045-b8q4q" Dec 03 00:45:00 crc kubenswrapper[4953]: I1203 00:45:00.354421 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/df482960-425e-4b6b-9284-cedcc918cc94-secret-volume\") pod \"collect-profiles-29412045-b8q4q\" (UID: \"df482960-425e-4b6b-9284-cedcc918cc94\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412045-b8q4q" Dec 03 00:45:00 crc kubenswrapper[4953]: I1203 00:45:00.359685 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5sj8c\" (UniqueName: \"kubernetes.io/projected/df482960-425e-4b6b-9284-cedcc918cc94-kube-api-access-5sj8c\") pod \"collect-profiles-29412045-b8q4q\" (UID: \"df482960-425e-4b6b-9284-cedcc918cc94\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412045-b8q4q" Dec 03 00:45:00 crc kubenswrapper[4953]: I1203 00:45:00.478622 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29412045-b8q4q" Dec 03 00:45:00 crc kubenswrapper[4953]: I1203 00:45:00.906892 4953 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412045-b8q4q"] Dec 03 00:45:01 crc kubenswrapper[4953]: I1203 00:45:01.021575 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29412045-b8q4q" event={"ID":"df482960-425e-4b6b-9284-cedcc918cc94","Type":"ContainerStarted","Data":"52fd943c5eb4538c9be1c7b0b4af67b0cfd5a983f2df704dbaaa46ca01a23daf"} Dec 03 00:45:02 crc kubenswrapper[4953]: I1203 00:45:02.031387 4953 generic.go:334] "Generic (PLEG): container finished" podID="df482960-425e-4b6b-9284-cedcc918cc94" containerID="631dd5402386ceb6534d000fe80024f664c80a1614874b703cbdd24fffbc2521" exitCode=0 Dec 03 00:45:02 crc kubenswrapper[4953]: I1203 00:45:02.031469 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29412045-b8q4q" event={"ID":"df482960-425e-4b6b-9284-cedcc918cc94","Type":"ContainerDied","Data":"631dd5402386ceb6534d000fe80024f664c80a1614874b703cbdd24fffbc2521"} Dec 03 00:45:03 crc kubenswrapper[4953]: I1203 00:45:03.345611 4953 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29412045-b8q4q" Dec 03 00:45:03 crc kubenswrapper[4953]: I1203 00:45:03.376783 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5sj8c\" (UniqueName: \"kubernetes.io/projected/df482960-425e-4b6b-9284-cedcc918cc94-kube-api-access-5sj8c\") pod \"df482960-425e-4b6b-9284-cedcc918cc94\" (UID: \"df482960-425e-4b6b-9284-cedcc918cc94\") " Dec 03 00:45:03 crc kubenswrapper[4953]: I1203 00:45:03.377116 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/df482960-425e-4b6b-9284-cedcc918cc94-secret-volume\") pod \"df482960-425e-4b6b-9284-cedcc918cc94\" (UID: \"df482960-425e-4b6b-9284-cedcc918cc94\") " Dec 03 00:45:03 crc kubenswrapper[4953]: I1203 00:45:03.377738 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/df482960-425e-4b6b-9284-cedcc918cc94-config-volume\") pod \"df482960-425e-4b6b-9284-cedcc918cc94\" (UID: \"df482960-425e-4b6b-9284-cedcc918cc94\") " Dec 03 00:45:03 crc kubenswrapper[4953]: I1203 00:45:03.378958 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/df482960-425e-4b6b-9284-cedcc918cc94-config-volume" (OuterVolumeSpecName: "config-volume") pod "df482960-425e-4b6b-9284-cedcc918cc94" (UID: "df482960-425e-4b6b-9284-cedcc918cc94"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:45:03 crc kubenswrapper[4953]: I1203 00:45:03.385230 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/df482960-425e-4b6b-9284-cedcc918cc94-kube-api-access-5sj8c" (OuterVolumeSpecName: "kube-api-access-5sj8c") pod "df482960-425e-4b6b-9284-cedcc918cc94" (UID: "df482960-425e-4b6b-9284-cedcc918cc94"). InnerVolumeSpecName "kube-api-access-5sj8c". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:45:03 crc kubenswrapper[4953]: I1203 00:45:03.385262 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/df482960-425e-4b6b-9284-cedcc918cc94-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "df482960-425e-4b6b-9284-cedcc918cc94" (UID: "df482960-425e-4b6b-9284-cedcc918cc94"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:45:03 crc kubenswrapper[4953]: I1203 00:45:03.480068 4953 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5sj8c\" (UniqueName: \"kubernetes.io/projected/df482960-425e-4b6b-9284-cedcc918cc94-kube-api-access-5sj8c\") on node \"crc\" DevicePath \"\"" Dec 03 00:45:03 crc kubenswrapper[4953]: I1203 00:45:03.480102 4953 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/df482960-425e-4b6b-9284-cedcc918cc94-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 03 00:45:03 crc kubenswrapper[4953]: I1203 00:45:03.480116 4953 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/df482960-425e-4b6b-9284-cedcc918cc94-config-volume\") on node \"crc\" DevicePath \"\"" Dec 03 00:45:04 crc kubenswrapper[4953]: I1203 00:45:04.054847 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29412045-b8q4q" event={"ID":"df482960-425e-4b6b-9284-cedcc918cc94","Type":"ContainerDied","Data":"52fd943c5eb4538c9be1c7b0b4af67b0cfd5a983f2df704dbaaa46ca01a23daf"} Dec 03 00:45:04 crc kubenswrapper[4953]: I1203 00:45:04.054921 4953 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="52fd943c5eb4538c9be1c7b0b4af67b0cfd5a983f2df704dbaaa46ca01a23daf" Dec 03 00:45:04 crc kubenswrapper[4953]: I1203 00:45:04.055007 4953 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29412045-b8q4q" Dec 03 00:45:04 crc kubenswrapper[4953]: I1203 00:45:04.461626 4953 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412000-mqgxv"] Dec 03 00:45:04 crc kubenswrapper[4953]: I1203 00:45:04.469209 4953 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412000-mqgxv"] Dec 03 00:45:05 crc kubenswrapper[4953]: I1203 00:45:05.366148 4953 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7b91526b-2492-4c17-949b-0ba3f8c221ea" path="/var/lib/kubelet/pods/7b91526b-2492-4c17-949b-0ba3f8c221ea/volumes" Dec 03 00:45:18 crc kubenswrapper[4953]: I1203 00:45:18.944670 4953 patch_prober.go:28] interesting pod/machine-config-daemon-6dfbn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 00:45:18 crc kubenswrapper[4953]: I1203 00:45:18.945676 4953 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-6dfbn" podUID="53ffb1e9-9a2e-4869-80d7-776f4aa7ef8b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 00:45:48 crc kubenswrapper[4953]: I1203 00:45:48.944718 4953 patch_prober.go:28] interesting pod/machine-config-daemon-6dfbn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 00:45:48 crc kubenswrapper[4953]: I1203 00:45:48.945502 4953 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-6dfbn" podUID="53ffb1e9-9a2e-4869-80d7-776f4aa7ef8b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 00:45:52 crc kubenswrapper[4953]: I1203 00:45:52.983574 4953 scope.go:117] "RemoveContainer" containerID="c9bb8e9e994fc6c10d6aab1c55ae333228332a4ed3369558948c3249ba7c71af" Dec 03 00:45:59 crc kubenswrapper[4953]: I1203 00:45:59.339333 4953 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["service-telemetry/infrawatch-operators-gsfkf"] Dec 03 00:45:59 crc kubenswrapper[4953]: E1203 00:45:59.340677 4953 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="df482960-425e-4b6b-9284-cedcc918cc94" containerName="collect-profiles" Dec 03 00:45:59 crc kubenswrapper[4953]: I1203 00:45:59.340700 4953 state_mem.go:107] "Deleted CPUSet assignment" podUID="df482960-425e-4b6b-9284-cedcc918cc94" containerName="collect-profiles" Dec 03 00:45:59 crc kubenswrapper[4953]: I1203 00:45:59.340907 4953 memory_manager.go:354] "RemoveStaleState removing state" podUID="df482960-425e-4b6b-9284-cedcc918cc94" containerName="collect-profiles" Dec 03 00:45:59 crc kubenswrapper[4953]: I1203 00:45:59.341771 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/infrawatch-operators-gsfkf" Dec 03 00:45:59 crc kubenswrapper[4953]: I1203 00:45:59.362610 4953 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/infrawatch-operators-gsfkf"] Dec 03 00:45:59 crc kubenswrapper[4953]: I1203 00:45:59.535203 4953 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qmwwq\" (UniqueName: \"kubernetes.io/projected/c667af35-f1ae-4f33-b2f5-9036059b6a11-kube-api-access-qmwwq\") pod \"infrawatch-operators-gsfkf\" (UID: \"c667af35-f1ae-4f33-b2f5-9036059b6a11\") " pod="service-telemetry/infrawatch-operators-gsfkf" Dec 03 00:45:59 crc kubenswrapper[4953]: I1203 00:45:59.636775 4953 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qmwwq\" (UniqueName: \"kubernetes.io/projected/c667af35-f1ae-4f33-b2f5-9036059b6a11-kube-api-access-qmwwq\") pod \"infrawatch-operators-gsfkf\" (UID: \"c667af35-f1ae-4f33-b2f5-9036059b6a11\") " pod="service-telemetry/infrawatch-operators-gsfkf" Dec 03 00:45:59 crc kubenswrapper[4953]: I1203 00:45:59.658081 4953 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qmwwq\" (UniqueName: \"kubernetes.io/projected/c667af35-f1ae-4f33-b2f5-9036059b6a11-kube-api-access-qmwwq\") pod \"infrawatch-operators-gsfkf\" (UID: \"c667af35-f1ae-4f33-b2f5-9036059b6a11\") " pod="service-telemetry/infrawatch-operators-gsfkf" Dec 03 00:45:59 crc kubenswrapper[4953]: I1203 00:45:59.665852 4953 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/infrawatch-operators-gsfkf" Dec 03 00:45:59 crc kubenswrapper[4953]: I1203 00:45:59.939894 4953 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/infrawatch-operators-gsfkf"] Dec 03 00:46:00 crc kubenswrapper[4953]: I1203 00:46:00.613321 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/infrawatch-operators-gsfkf" event={"ID":"c667af35-f1ae-4f33-b2f5-9036059b6a11","Type":"ContainerStarted","Data":"b23fa4e26708310178ba160fdbfcbce34bb131aacf752efcb1ee25afef58a576"} Dec 03 00:46:00 crc kubenswrapper[4953]: I1203 00:46:00.613777 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/infrawatch-operators-gsfkf" event={"ID":"c667af35-f1ae-4f33-b2f5-9036059b6a11","Type":"ContainerStarted","Data":"dfda51e6c8086760eb18188df05eedb6bc959867a950a5af5fca825640948712"} Dec 03 00:46:00 crc kubenswrapper[4953]: I1203 00:46:00.639487 4953 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="service-telemetry/infrawatch-operators-gsfkf" podStartSLOduration=1.238605406 podStartE2EDuration="1.639466708s" podCreationTimestamp="2025-12-03 00:45:59 +0000 UTC" firstStartedPulling="2025-12-03 00:45:59.955380426 +0000 UTC m=+2348.939661219" lastFinishedPulling="2025-12-03 00:46:00.356241728 +0000 UTC m=+2349.340522521" observedRunningTime="2025-12-03 00:46:00.633861749 +0000 UTC m=+2349.618142552" watchObservedRunningTime="2025-12-03 00:46:00.639466708 +0000 UTC m=+2349.623747491" Dec 03 00:46:09 crc kubenswrapper[4953]: I1203 00:46:09.666874 4953 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="service-telemetry/infrawatch-operators-gsfkf" Dec 03 00:46:09 crc kubenswrapper[4953]: I1203 00:46:09.668669 4953 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="service-telemetry/infrawatch-operators-gsfkf" Dec 03 00:46:09 crc kubenswrapper[4953]: I1203 00:46:09.706899 4953 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="service-telemetry/infrawatch-operators-gsfkf" Dec 03 00:46:09 crc kubenswrapper[4953]: I1203 00:46:09.745170 4953 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="service-telemetry/infrawatch-operators-gsfkf" Dec 03 00:46:09 crc kubenswrapper[4953]: I1203 00:46:09.798580 4953 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["service-telemetry/infrawatch-operators-gsfkf"] Dec 03 00:46:11 crc kubenswrapper[4953]: I1203 00:46:11.702093 4953 kuberuntime_container.go:808] "Killing container with a grace period" pod="service-telemetry/infrawatch-operators-gsfkf" podUID="c667af35-f1ae-4f33-b2f5-9036059b6a11" containerName="registry-server" containerID="cri-o://b23fa4e26708310178ba160fdbfcbce34bb131aacf752efcb1ee25afef58a576" gracePeriod=2 Dec 03 00:46:12 crc kubenswrapper[4953]: I1203 00:46:12.189306 4953 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/infrawatch-operators-gsfkf" Dec 03 00:46:12 crc kubenswrapper[4953]: I1203 00:46:12.342592 4953 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qmwwq\" (UniqueName: \"kubernetes.io/projected/c667af35-f1ae-4f33-b2f5-9036059b6a11-kube-api-access-qmwwq\") pod \"c667af35-f1ae-4f33-b2f5-9036059b6a11\" (UID: \"c667af35-f1ae-4f33-b2f5-9036059b6a11\") " Dec 03 00:46:12 crc kubenswrapper[4953]: I1203 00:46:12.356344 4953 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c667af35-f1ae-4f33-b2f5-9036059b6a11-kube-api-access-qmwwq" (OuterVolumeSpecName: "kube-api-access-qmwwq") pod "c667af35-f1ae-4f33-b2f5-9036059b6a11" (UID: "c667af35-f1ae-4f33-b2f5-9036059b6a11"). InnerVolumeSpecName "kube-api-access-qmwwq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:46:12 crc kubenswrapper[4953]: I1203 00:46:12.447304 4953 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qmwwq\" (UniqueName: \"kubernetes.io/projected/c667af35-f1ae-4f33-b2f5-9036059b6a11-kube-api-access-qmwwq\") on node \"crc\" DevicePath \"\"" Dec 03 00:46:12 crc kubenswrapper[4953]: I1203 00:46:12.715915 4953 generic.go:334] "Generic (PLEG): container finished" podID="c667af35-f1ae-4f33-b2f5-9036059b6a11" containerID="b23fa4e26708310178ba160fdbfcbce34bb131aacf752efcb1ee25afef58a576" exitCode=0 Dec 03 00:46:12 crc kubenswrapper[4953]: I1203 00:46:12.715991 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/infrawatch-operators-gsfkf" event={"ID":"c667af35-f1ae-4f33-b2f5-9036059b6a11","Type":"ContainerDied","Data":"b23fa4e26708310178ba160fdbfcbce34bb131aacf752efcb1ee25afef58a576"} Dec 03 00:46:12 crc kubenswrapper[4953]: I1203 00:46:12.716030 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/infrawatch-operators-gsfkf" event={"ID":"c667af35-f1ae-4f33-b2f5-9036059b6a11","Type":"ContainerDied","Data":"dfda51e6c8086760eb18188df05eedb6bc959867a950a5af5fca825640948712"} Dec 03 00:46:12 crc kubenswrapper[4953]: I1203 00:46:12.716050 4953 scope.go:117] "RemoveContainer" containerID="b23fa4e26708310178ba160fdbfcbce34bb131aacf752efcb1ee25afef58a576" Dec 03 00:46:12 crc kubenswrapper[4953]: I1203 00:46:12.716292 4953 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/infrawatch-operators-gsfkf" Dec 03 00:46:12 crc kubenswrapper[4953]: I1203 00:46:12.753644 4953 scope.go:117] "RemoveContainer" containerID="b23fa4e26708310178ba160fdbfcbce34bb131aacf752efcb1ee25afef58a576" Dec 03 00:46:12 crc kubenswrapper[4953]: E1203 00:46:12.754344 4953 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b23fa4e26708310178ba160fdbfcbce34bb131aacf752efcb1ee25afef58a576\": container with ID starting with b23fa4e26708310178ba160fdbfcbce34bb131aacf752efcb1ee25afef58a576 not found: ID does not exist" containerID="b23fa4e26708310178ba160fdbfcbce34bb131aacf752efcb1ee25afef58a576" Dec 03 00:46:12 crc kubenswrapper[4953]: I1203 00:46:12.754393 4953 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b23fa4e26708310178ba160fdbfcbce34bb131aacf752efcb1ee25afef58a576"} err="failed to get container status \"b23fa4e26708310178ba160fdbfcbce34bb131aacf752efcb1ee25afef58a576\": rpc error: code = NotFound desc = could not find container \"b23fa4e26708310178ba160fdbfcbce34bb131aacf752efcb1ee25afef58a576\": container with ID starting with b23fa4e26708310178ba160fdbfcbce34bb131aacf752efcb1ee25afef58a576 not found: ID does not exist" Dec 03 00:46:12 crc kubenswrapper[4953]: I1203 00:46:12.771581 4953 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["service-telemetry/infrawatch-operators-gsfkf"] Dec 03 00:46:12 crc kubenswrapper[4953]: I1203 00:46:12.779857 4953 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["service-telemetry/infrawatch-operators-gsfkf"] Dec 03 00:46:13 crc kubenswrapper[4953]: I1203 00:46:13.356342 4953 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c667af35-f1ae-4f33-b2f5-9036059b6a11" path="/var/lib/kubelet/pods/c667af35-f1ae-4f33-b2f5-9036059b6a11/volumes" Dec 03 00:46:18 crc kubenswrapper[4953]: I1203 00:46:18.945554 4953 patch_prober.go:28] interesting pod/machine-config-daemon-6dfbn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 00:46:18 crc kubenswrapper[4953]: I1203 00:46:18.946313 4953 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-6dfbn" podUID="53ffb1e9-9a2e-4869-80d7-776f4aa7ef8b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 00:46:18 crc kubenswrapper[4953]: I1203 00:46:18.946381 4953 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-6dfbn" Dec 03 00:46:18 crc kubenswrapper[4953]: I1203 00:46:18.947205 4953 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"3aa2983089e4009bcb4ec8782554c6ba514ea720700589b3adbf3c56a194c241"} pod="openshift-machine-config-operator/machine-config-daemon-6dfbn" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 03 00:46:18 crc kubenswrapper[4953]: I1203 00:46:18.947293 4953 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-6dfbn" podUID="53ffb1e9-9a2e-4869-80d7-776f4aa7ef8b" containerName="machine-config-daemon" containerID="cri-o://3aa2983089e4009bcb4ec8782554c6ba514ea720700589b3adbf3c56a194c241" gracePeriod=600 Dec 03 00:46:19 crc kubenswrapper[4953]: E1203 00:46:19.077119 4953 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6dfbn_openshift-machine-config-operator(53ffb1e9-9a2e-4869-80d7-776f4aa7ef8b)\"" pod="openshift-machine-config-operator/machine-config-daemon-6dfbn" podUID="53ffb1e9-9a2e-4869-80d7-776f4aa7ef8b" Dec 03 00:46:19 crc kubenswrapper[4953]: I1203 00:46:19.789128 4953 generic.go:334] "Generic (PLEG): container finished" podID="53ffb1e9-9a2e-4869-80d7-776f4aa7ef8b" containerID="3aa2983089e4009bcb4ec8782554c6ba514ea720700589b3adbf3c56a194c241" exitCode=0 Dec 03 00:46:19 crc kubenswrapper[4953]: I1203 00:46:19.789263 4953 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-6dfbn" event={"ID":"53ffb1e9-9a2e-4869-80d7-776f4aa7ef8b","Type":"ContainerDied","Data":"3aa2983089e4009bcb4ec8782554c6ba514ea720700589b3adbf3c56a194c241"} Dec 03 00:46:19 crc kubenswrapper[4953]: I1203 00:46:19.789536 4953 scope.go:117] "RemoveContainer" containerID="e9bb3f6087811f1ea895def278b13c2a3d71dcdf55ad22dee5da1051917891f2" Dec 03 00:46:19 crc kubenswrapper[4953]: I1203 00:46:19.790379 4953 scope.go:117] "RemoveContainer" containerID="3aa2983089e4009bcb4ec8782554c6ba514ea720700589b3adbf3c56a194c241" Dec 03 00:46:19 crc kubenswrapper[4953]: E1203 00:46:19.790768 4953 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6dfbn_openshift-machine-config-operator(53ffb1e9-9a2e-4869-80d7-776f4aa7ef8b)\"" pod="openshift-machine-config-operator/machine-config-daemon-6dfbn" podUID="53ffb1e9-9a2e-4869-80d7-776f4aa7ef8b" Dec 03 00:46:34 crc kubenswrapper[4953]: I1203 00:46:34.349193 4953 scope.go:117] "RemoveContainer" containerID="3aa2983089e4009bcb4ec8782554c6ba514ea720700589b3adbf3c56a194c241" Dec 03 00:46:34 crc kubenswrapper[4953]: E1203 00:46:34.350838 4953 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6dfbn_openshift-machine-config-operator(53ffb1e9-9a2e-4869-80d7-776f4aa7ef8b)\"" pod="openshift-machine-config-operator/machine-config-daemon-6dfbn" podUID="53ffb1e9-9a2e-4869-80d7-776f4aa7ef8b" Dec 03 00:46:47 crc kubenswrapper[4953]: I1203 00:46:47.348397 4953 scope.go:117] "RemoveContainer" containerID="3aa2983089e4009bcb4ec8782554c6ba514ea720700589b3adbf3c56a194c241" Dec 03 00:46:47 crc kubenswrapper[4953]: E1203 00:46:47.349733 4953 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6dfbn_openshift-machine-config-operator(53ffb1e9-9a2e-4869-80d7-776f4aa7ef8b)\"" pod="openshift-machine-config-operator/machine-config-daemon-6dfbn" podUID="53ffb1e9-9a2e-4869-80d7-776f4aa7ef8b" Dec 03 00:47:00 crc kubenswrapper[4953]: I1203 00:47:00.349222 4953 scope.go:117] "RemoveContainer" containerID="3aa2983089e4009bcb4ec8782554c6ba514ea720700589b3adbf3c56a194c241" Dec 03 00:47:00 crc kubenswrapper[4953]: E1203 00:47:00.350200 4953 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-6dfbn_openshift-machine-config-operator(53ffb1e9-9a2e-4869-80d7-776f4aa7ef8b)\"" pod="openshift-machine-config-operator/machine-config-daemon-6dfbn" podUID="53ffb1e9-9a2e-4869-80d7-776f4aa7ef8b" var/home/core/zuul-output/logs/crc-cloud-workdir-crc-all-logs.tar.gz0000644000175000000000000000005515113704231024442 0ustar coreroot‹íÁ  ÷Om7 €7šÞ'(var/home/core/zuul-output/logs/crc-cloud/0000755000175000000000000000000015113704232017360 5ustar corerootvar/home/core/zuul-output/artifacts/0000755000175000017500000000000015113677113016512 5ustar corecorevar/home/core/zuul-output/docs/0000755000175000017500000000000015113677114015463 5ustar corecore